Hadoop not utilizing max reducer capacity + reducer stuck in pending state

2010-08-18 Thread Tarandeep Singh
Hi, I am seeing some strange behavior in Hadoop - I am running a small test cluster with a capacity of 18 mappers and 18 reducers. I fire a lot of jobs simultaneously and over time I have observed Hadoop is not utilizing all the 18 slots for the reducers. And now even if I run just one job (no

Re: Hadoop not utilizing max reducer capacity + reducer stuck in pending state

2010-08-18 Thread Tarandeep Singh
other reducers had finished. On Wed, Aug 18, 2010 at 2:44 PM, Tarandeep Singh tarand...@gmail.comwrote: Hi, I am seeing some strange behavior in Hadoop - I am running a small test cluster with a capacity of 18 mappers and 18 reducers. I fire a lot of jobs simultaneously and over time I have

Re: Hadoop and X11 related error

2010-01-18 Thread Tarandeep Singh
On Mon, Jan 18, 2010 at 2:52 AM, Steve Loughran ste...@apache.org wrote: Tarandeep Singh wrote: Hi, I am running a MR job that requires usage of some java.awt.* classes, that can't be run in headless mode. Right now, I am running Hadoop in a single node cluster (my laptop) which has X11

Hadoop and X11 related error

2010-01-17 Thread Tarandeep Singh
Hi, I am running a MR job that requires usage of some java.awt.* classes, that can't be run in headless mode. Right now, I am running Hadoop in a single node cluster (my laptop) which has X11 server running. I have set up my ssh server and client to do X11 forwarding. I ran the following java

Re: Hadoop and X11 related error

2010-01-17 Thread Tarandeep Singh
either some permission issues or what) On Jan 18, 2010, at 12:41 AM, Tarandeep Singh wrote: Hi, I am running a MR job that requires usage of some java.awt.* classes, that can't be run in headless mode. Right now, I am running Hadoop in a single node cluster (my laptop) which has

Re: How to retrieve the reducer output file names?

2009-09-12 Thread Tarandeep Singh
The output of mappers is partitioned, each partition is given a number starting from 0 and a reducer works on one of these partitions. In the configure method of your reducer code, you can get the partition number by- jobConf.getInt( mapred.task.partition, 0); If you use the default output

Re: Doubt regarding Replication Factor

2009-08-12 Thread Tarandeep Singh
A similar question- If in an N node cluster, a file's replication is set to N (replicate on each node) and later if a node goes down, will HDFS throw an exception since the file's replication has gone down below the specified number ? Thanks, Tarandeep On Wed, Aug 12, 2009 at 12:11 PM,

Re: Compression issues!!

2009-07-15 Thread Tarandeep Singh
You can put compress data on HDFS and run Map Reduce job on it. But you should use a codec that supports file splitting, otherwise whole file will be read by one mapper. If you have read about Map reduce architecture, you would understand that a map function processes chunk of data (called split).