On Apr 23, 2008, at 7:51 AM, Apurva Jadhav wrote:

There are six reducers and 24000 mappers because there are 24000 files.
The number of tasks per node is 2.
mapred.child.java opts is the default value 200m. What is a good value for this.? My mappers and reducers are fairly simple and do not make large allocations.

Try upping that to 512M.

Arun

Regards,
aj

Amar Kamat wrote:
Apurva Jadhav wrote:
Hi,
I have a 4 node hadoop 0.15.3 cluster. I am using the default config files. I am running a map reduce job to process 40 GB log data.
How many maps and reducers are there? Make sure that there are sufficient number of reducers. Look at conf/hadoop-default.xml (see mapred.child.java.opts parameter) to change the heap settings.
Amar
Some reduce tasks are failing with the following errors:
1)
stderr
Exception in thread "org.apache.hadoop.io.ObjectWritable Connection Culler" Exception in thread "[EMAIL PROTECTED]" java.lang.OutOfMemoryError: Java heap space Exception in thread "IPC Client connection to /127.0.0.1:34691" java.lang.OutOfMemoryError: Java heap space Exception in thread "main" java.lang.OutOfMemoryError: Java heap space

2)
stderr
Exception in thread "org.apache.hadoop.io.ObjectWritable Connection Culler" java.lang.OutOfMemoryError: Java heap space

syslog:
2008-04-22 19:32:50,784 INFO org.apache.hadoop.mapred.ReduceTask: task_200804212359_0007_r_000004_0 Merge of the 19 files in InMemoryFileSystem complete. Local file is /data/hadoop-im2/ mapred/loca
l/task_200804212359_0007_r_000004_0/map_22600.out
2008-04-22 20:34:16,012 INFO org.apache.hadoop.ipc.Client: java.net.SocketException: Socket closed at java.net.SocketInputStream.read(SocketInputStream.java: 162)
       at java.io.FilterInputStream.read(FilterInputStream.java:111)
at org.apache.hadoop.ipc.Client$Connection$1.read (Client.java:181) at java.io.BufferedInputStream.fill (BufferedInputStream.java:218) at java.io.BufferedInputStream.read (BufferedInputStream.java:235)
       at java.io.DataInputStream.readInt(DataInputStream.java:353)
at org.apache.hadoop.ipc.Client$Connection.run(Client.java: 258)

2008-04-22 20:34:16,032 WARN org.apache.hadoop.mapred.TaskTracker: Error running child
java.lang.OutOfMemoryError: Java heap space
2008-04-22 20:34:16,031 INFO org.apache.hadoop.mapred.TaskRunner: Communication exception: java.lang.OutOfMemoryError: Java heap space

Has anyone experienced similar problem ? Is there any configuration change that can help resolve this issue.

Regards,
aj







Reply via email to