Thanks, I tried but this probably not the reason. I checked the network
connection using "netstat" and the client is actually connected to the
correct server address. In addition, "mrbench" works sometime; if it is network problem, nothing should work at all.

I let the "sort" program ran longer, and get some interesting output, the
reduce progress can actually be descreasing and oscillating (see the bottom part of the output below).

The error information:
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
08/01/22 13:56:21 WARN mapred.JobClient: Error reading task outputncdm15

What can cause "Error reading task outputncdm15"?

[EMAIL PROTECTED]:~/hadoop-0.15.2$ ./bin/hadoop jar hadoop-0.15.2-examples.jar 
sort rand randsort
Running on 2 nodes to sort from /user/gu/rand into /user/gu/randsort with 8 
reduces.
Job started: Tue Jan 22 11:56:39 CST 2008
08/01/22 11:56:39 INFO mapred.FileInputFormat: Total input paths to process: 20
08/01/22 11:56:40 INFO mapred.JobClient: Running job: job_200801221154_0005
08/01/22 11:56:41 INFO mapred.JobClient:  map 0% reduce 0%
08/01/22 11:56:54 INFO mapred.JobClient:  map 1% reduce 0%
08/01/22 11:56:58 INFO mapred.JobClient:  map 2% reduce 0%
08/01/22 11:57:04 INFO mapred.JobClient:  map 3% reduce 0%
08/01/22 11:57:08 INFO mapred.JobClient:  map 5% reduce 0%
08/01/22 11:57:15 INFO mapred.JobClient:  map 7% reduce 0%
08/01/22 11:57:23 INFO mapred.JobClient:  map 8% reduce 0%
08/01/22 11:57:25 INFO mapred.JobClient:  map 9% reduce 0%
08/01/22 11:57:27 INFO mapred.JobClient:  map 10% reduce 0%
08/01/22 11:57:34 INFO mapred.JobClient:  map 11% reduce 0%
08/01/22 11:57:35 INFO mapred.JobClient:  map 12% reduce 0%
08/01/22 11:57:42 INFO mapred.JobClient:  map 13% reduce 0%
08/01/22 11:57:48 INFO mapred.JobClient:  map 14% reduce 0%
08/01/22 11:57:49 INFO mapred.JobClient:  map 15% reduce 0%
08/01/22 11:57:56 INFO mapred.JobClient:  map 17% reduce 0%
08/01/22 11:58:05 INFO mapred.JobClient:  map 19% reduce 0%
08/01/22 11:58:12 INFO mapred.JobClient:  map 20% reduce 0%
08/01/22 11:58:13 INFO mapred.JobClient:  map 20% reduce 1%
08/01/22 11:58:14 INFO mapred.JobClient:  map 21% reduce 1%
08/01/22 11:58:19 INFO mapred.JobClient:  map 22% reduce 1%
08/01/22 11:58:26 INFO mapred.JobClient:  map 23% reduce 1%
08/01/22 11:58:30 INFO mapred.JobClient:  map 24% reduce 1%
08/01/22 11:58:32 INFO mapred.JobClient:  map 25% reduce 1%
08/01/22 11:58:37 INFO mapred.JobClient:  map 26% reduce 1%
08/01/22 11:58:42 INFO mapred.JobClient:  map 27% reduce 1%
08/01/22 11:58:44 INFO mapred.JobClient:  map 28% reduce 1%
08/01/22 11:58:49 INFO mapred.JobClient:  map 29% reduce 1%
08/01/22 11:58:55 INFO mapred.JobClient:  map 30% reduce 1%
08/01/22 11:58:57 INFO mapred.JobClient:  map 31% reduce 1%
08/01/22 11:58:59 INFO mapred.JobClient:  map 32% reduce 1%
08/01/22 11:59:07 INFO mapred.JobClient:  map 33% reduce 1%
08/01/22 11:59:08 INFO mapred.JobClient:  map 34% reduce 1%
08/01/22 11:59:15 INFO mapred.JobClient:  map 35% reduce 1%
08/01/22 11:59:17 INFO mapred.JobClient:  map 36% reduce 2%
08/01/22 11:59:20 INFO mapred.JobClient:  map 37% reduce 2%
08/01/22 11:59:27 INFO mapred.JobClient:  map 38% reduce 2%
08/01/22 11:59:29 INFO mapred.JobClient:  map 39% reduce 2%
08/01/22 11:59:34 INFO mapred.JobClient:  map 40% reduce 2%
08/01/22 11:59:38 INFO mapred.JobClient:  map 41% reduce 2%
08/01/22 11:59:41 INFO mapred.JobClient:  map 42% reduce 2%
08/01/22 11:59:42 INFO mapred.JobClient:  map 43% reduce 2%
08/01/22 11:59:51 INFO mapred.JobClient:  map 44% reduce 2%
08/01/22 11:59:57 INFO mapred.JobClient:  map 46% reduce 2%
08/01/22 12:00:01 INFO mapred.JobClient:  map 47% reduce 2%
08/01/22 12:00:06 INFO mapred.JobClient:  map 48% reduce 2%
08/01/22 12:00:08 INFO mapred.JobClient:  map 49% reduce 2%
08/01/22 12:00:14 INFO mapred.JobClient:  map 50% reduce 2%
08/01/22 12:00:16 INFO mapred.JobClient:  map 51% reduce 2%
08/01/22 12:00:22 INFO mapred.JobClient:  map 52% reduce 2%
08/01/22 12:00:27 INFO mapred.JobClient:  map 53% reduce 3%
08/01/22 12:00:30 INFO mapred.JobClient:  map 54% reduce 3%
08/01/22 12:00:32 INFO mapred.JobClient:  map 55% reduce 3%
08/01/22 12:00:37 INFO mapred.JobClient:  map 56% reduce 3%
08/01/22 12:00:38 INFO mapred.JobClient:  map 57% reduce 3%
08/01/22 12:00:44 INFO mapred.JobClient:  map 58% reduce 3%
08/01/22 12:00:49 INFO mapred.JobClient:  map 59% reduce 3%
08/01/22 12:00:52 INFO mapred.JobClient:  map 60% reduce 3%
08/01/22 12:00:58 INFO mapred.JobClient:  map 61% reduce 3%
08/01/22 12:00:59 INFO mapred.JobClient:  map 62% reduce 3%
08/01/22 12:01:05 INFO mapred.JobClient:  map 63% reduce 3%
08/01/22 12:01:08 INFO mapred.JobClient:  map 64% reduce 3%
08/01/22 12:01:13 INFO mapred.JobClient:  map 65% reduce 3%
08/01/22 12:01:14 INFO mapred.JobClient:  map 66% reduce 3%
08/01/22 12:01:20 INFO mapred.JobClient:  map 67% reduce 3%
08/01/22 12:01:25 INFO mapred.JobClient:  map 68% reduce 3%
08/01/22 12:01:28 INFO mapred.JobClient:  map 69% reduce 3%
08/01/22 12:01:35 INFO mapred.JobClient:  map 71% reduce 4%
08/01/22 12:01:39 INFO mapred.JobClient:  map 72% reduce 4%
08/01/22 12:01:46 INFO mapred.JobClient:  map 74% reduce 4%
08/01/22 12:01:51 INFO mapred.JobClient:  map 75% reduce 4%
08/01/22 12:01:54 INFO mapred.JobClient:  map 76% reduce 4%
08/01/22 12:02:01 INFO mapred.JobClient:  map 77% reduce 4%
08/01/22 12:02:04 INFO mapred.JobClient:  map 78% reduce 4%
08/01/22 12:02:10 INFO mapred.JobClient:  map 79% reduce 4%
08/01/22 12:02:13 INFO mapred.JobClient:  map 80% reduce 4%
08/01/22 12:02:14 INFO mapred.JobClient:  map 81% reduce 4%
08/01/22 12:02:17 INFO mapred.JobClient:  map 82% reduce 4%
08/01/22 12:02:21 INFO mapred.JobClient:  map 83% reduce 4%
08/01/22 12:02:29 INFO mapred.JobClient:  map 84% reduce 4%
08/01/22 12:02:31 INFO mapred.JobClient:  map 85% reduce 4%
08/01/22 12:02:36 INFO mapred.JobClient:  map 86% reduce 4%
08/01/22 12:02:39 INFO mapred.JobClient:  map 87% reduce 4%
08/01/22 12:02:43 INFO mapred.JobClient:  map 88% reduce 4%
08/01/22 12:02:46 INFO mapred.JobClient:  map 89% reduce 4%
08/01/22 12:02:50 INFO mapred.JobClient:  map 90% reduce 4%
08/01/22 12:02:56 INFO mapred.JobClient:  map 91% reduce 5%
08/01/22 12:02:58 INFO mapred.JobClient:  map 92% reduce 5%
08/01/22 12:03:05 INFO mapred.JobClient:  map 93% reduce 5%
08/01/22 12:03:07 INFO mapred.JobClient:  map 94% reduce 5%
08/01/22 12:03:12 INFO mapred.JobClient:  map 96% reduce 5%
08/01/22 12:03:20 INFO mapred.JobClient:  map 97% reduce 5%
08/01/22 12:03:24 INFO mapred.JobClient:  map 98% reduce 5%
08/01/22 12:03:28 INFO mapred.JobClient:  map 99% reduce 5%
08/01/22 12:03:35 INFO mapred.JobClient:  map 100% reduce 5%
08/01/22 12:03:45 INFO mapred.JobClient:  map 100% reduce 6%
08/01/22 12:04:05 INFO mapred.JobClient:  map 100% reduce 7%
08/01/22 12:04:25 INFO mapred.JobClient:  map 100% reduce 8%
08/01/22 12:04:52 INFO mapred.JobClient:  map 100% reduce 9%
08/01/22 12:05:15 INFO mapred.JobClient:  map 100% reduce 10%
08/01/22 12:05:43 INFO mapred.JobClient:  map 100% reduce 11%
08/01/22 12:06:13 INFO mapred.JobClient:  map 100% reduce 12%
08/01/22 12:06:32 INFO mapred.JobClient:  map 100% reduce 13%
08/01/22 12:07:02 INFO mapred.JobClient:  map 100% reduce 14%
08/01/22 12:07:25 INFO mapred.JobClient:  map 100% reduce 15%
08/01/22 12:07:52 INFO mapred.JobClient:  map 100% reduce 16%
08/01/22 13:56:21 INFO mapred.JobClient:  map 100% reduce 13%
08/01/22 13:56:21 INFO mapred.JobClient: Task Id :
task_200801221154_0005_r_000007_0, Status : FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
08/01/22 13:56:21 WARN mapred.JobClient: Error reading task outputncdm15
08/01/22 13:56:21 WARN mapred.JobClient: Error reading task outputncdm15
08/01/22 14:08:11 INFO mapred.JobClient:  map 100% reduce 11%
08/01/22 14:08:11 INFO mapred.JobClient: Task Id :
task_200801221154_0005_r_000005_0, Status : FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
08/01/22 14:08:11 WARN mapred.JobClient: Error reading task outputncdm15
08/01/22 14:08:11 WARN mapred.JobClient: Error reading task outputncdm15
08/01/22 14:20:09 INFO mapred.JobClient:  map 100% reduce 8%
08/01/22 14:20:09 INFO mapred.JobClient: Task Id :
task_200801221154_0005_r_000004_0, Status : FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
08/01/22 14:20:09 WARN mapred.JobClient: Error reading task outputncdm15
08/01/22 14:20:09 WARN mapred.JobClient: Error reading task outputncdm15
08/01/22 14:23:49 INFO mapred.JobClient:  map 100% reduce 5%
08/01/22 14:23:49 INFO mapred.JobClient: Task Id :
task_200801221154_0005_r_000006_0, Status : FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
08/01/22 14:23:49 WARN mapred.JobClient: Error reading task outputncdm15
08/01/22 14:23:49 WARN mapred.JobClient: Error reading task outputncdm15
08/01/22 15:10:19 INFO mapred.JobClient:  map 100% reduce 4%
08/01/22 15:10:19 INFO mapred.JobClient: Task Id :
task_200801221154_0005_r_000000_0, Status : FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
08/01/22 15:10:42 INFO mapred.JobClient:  map 100% reduce 5%
08/01/22 15:10:59 INFO mapred.JobClient:  map 100% reduce 6%
08/01/22 15:11:22 INFO mapred.JobClient:  map 100% reduce 7%
08/01/22 15:11:55 INFO mapred.JobClient:  map 100% reduce 8%
08/01/22 15:13:39 INFO mapred.JobClient:  map 100% reduce 7%
08/01/22 15:13:39 INFO mapred.JobClient: Task Id :
task_200801221154_0005_r_000003_0, Status : FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
08/01/22 15:14:09 INFO mapred.JobClient:  map 100% reduce 8%
08/01/22 15:14:29 INFO mapred.JobClient:  map 100% reduce 9%
08/01/22 15:14:52 INFO mapred.JobClient:  map 100% reduce 10%
08/01/22 15:15:33 INFO mapred.JobClient:  map 100% reduce 11%
08/01/22 15:37:01 INFO mapred.JobClient:  map 100% reduce 9%
08/01/22 15:37:01 INFO mapred.JobClient: Task Id :
task_200801221154_0005_r_000002_0, Status : FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
08/01/22 15:37:13 INFO mapred.JobClient:  map 100% reduce 10%
08/01/22 15:37:33 INFO mapred.JobClient:  map 100% reduce 11%
08/01/22 15:37:53 INFO mapred.JobClient:  map 100% reduce 12%
08/01/22 15:38:23 INFO mapred.JobClient:  map 100% reduce 13%
08/01/22 15:45:51 INFO mapred.JobClient:  map 100% reduce 12%
08/01/22 15:45:51 INFO mapred.JobClient: Task Id :
task_200801221154_0005_r_000001_0, Status : FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
08/01/22 15:46:11 INFO mapred.JobClient:  map 100% reduce 13%
08/01/22 15:46:31 INFO mapred.JobClient:  map 100% reduce 14%
08/01/22 15:46:51 INFO mapred.JobClient:  map 100% reduce 15%
08/01/22 15:47:24 INFO mapred.JobClient:  map 100% reduce 16%


On Tue, 22 Jan 2008, Taeho Kang wrote:

You said you had two network interfaces... and it might be the source of
your problem.

Try disabling one of your network interfaces
Or set "dfs.datanode.dns.interface" and "dfs.datanode.dns.nameserver" in
your config file so that datanode knows which network interface to pick up.

/Taeho

On Jan 22, 2008 7:34 AM, Yunhong Gu1 <[EMAIL PROTECTED]> wrote:


Hi, all

Just to keep this topic updated :) I am still tring to figure what
happened.

I found that in my 2-node configuration (namenode and jobtracker on
node-1, while both are datanodes and tasktrackers). The reduce task may
sometimes (but rarely) complete for programs that needs small amount of
CPU time (e.g., mrbench), but for programs with large computation, it
never finish. When reduces blocks, it always fails at 16%.

Eventually I will get this error information:
08/01/18 15:01:27 WARN mapred.JobClient: Error reading task
outputncdm-IPxxxxxxxxx
08/01/18 15:01:27 WARN mapred.JobClient: Error reading task
outputncdm-IPxxxxxxxxx
08/01/18 15:13:38 INFO mapred.JobClient: Task Id :
task_200801181145_0005_r_000000_1, Status : FAILED
Shuffle Error: Exceeded MAX_FAILED_UNIQUE_FETCHES; bailing-out.
08/01/18 19:56:38 WARN mapred.JobClient: Error reading task
outputConnection timed out
08/01/18 19:59:47 WARN mapred.JobClient: Error reading task
outputConnection timed out
08/01/18 20:09:40 INFO mapred.JobClient:  map 100% reduce 100%
java.io.IOException: Job failed!

I found that "IPxxxxxxxx" is not the correct network address
that Hadoop should read result from. The servers I use have 2 network
interfaces and I am using another one. I explicitly fill the IP addresses
10.0.0.x in all the configuration files.

Might this be the reason of Reduce failure? But the Map phase does work.

Thanks
Yunhong


On Sat, 19 Jan 2008, Yunhong Gu1 wrote:


Oh, so it is the task running on the other node (ncdm-15) fails and
Hadoop
re-run the task on the local node (ncdm-8). (I only have two nodes,
ncdm-8
and ncdm-15. Both namenode and jobtracker are running on ncdm-8. The
program
is also started on ncdm-8).

08/01/18 19:08:27 INFO mapred.JobClient: Task Id :
task_200801181852_0001_m_000001_0, Status : FAILED Too many
fetch-failures
08/01/18 19:08:27 WARN mapred.JobClient: Error reading task
outputncdm15

Any ideas why the task would fail? And why it takes so long for Hadoop
to
detect the failure?

Thanks
Yunhong

On Sat, 19 Jan 2008, Devaraj Das wrote:

Hi Yunhong,
As per the output it seems the job ran to successful completion (albeit

with
some failures)...
Devaraj

-----Original Message-----
From: Yunhong Gu1 [mailto:[EMAIL PROTECTED]
Sent: Saturday, January 19, 2008 8:56 AM
To: [EMAIL PROTECTED]
Subject: Re: Reduce hangs



Yes, it looks like HADOOP-1374

The program actually failed after a while:


[EMAIL PROTECTED]:~/hadoop-0.15.2$ ./bin/hadoop jar
hadoop-0.15.2-test.jar mrbench
MRBenchmark.0.0.2
08/01/18 18:53:08 INFO mapred.MRBench: creating control file:
1 numLines, ASCENDING sortOrder
08/01/18 18:53:08 INFO mapred.MRBench: created control file:
/benchmarks/MRBench/mr_input/input_-450753747.txt
08/01/18 18:53:09 INFO mapred.MRBench: Running job 0:
input=/benchmarks/MRBench/mr_input
output=/benchmarks/MRBench/mr_output/output_1843693325
08/01/18 18:53:09 INFO mapred.FileInputFormat: Total input
paths to process : 1
08/01/18 18:53:09 INFO mapred.JobClient: Running job:
job_200801181852_0001
08/01/18 18:53:10 INFO mapred.JobClient:  map 0% reduce 0%
08/01/18 18:53:17 INFO mapred.JobClient:  map 100% reduce 0%
08/01/18 18:53:25 INFO mapred.JobClient:  map 100% reduce 16%
08/01/18 19:08:27 INFO mapred.JobClient: Task Id :
task_200801181852_0001_m_000001_0, Status : FAILED Too many
fetch-failures
08/01/18 19:08:27 WARN mapred.JobClient: Error reading task
outputncdm15
08/01/18 19:08:27 WARN mapred.JobClient: Error reading task
outputncdm15
08/01/18 19:08:34 INFO mapred.JobClient:  map 100% reduce 100%
08/01/18 19:08:35 INFO mapred.JobClient: Job complete:
job_200801181852_0001
08/01/18 19:08:35 INFO mapred.JobClient: Counters: 10
08/01/18 19:08:35 INFO mapred.JobClient:   Job Counters
08/01/18 19:08:35 INFO mapred.JobClient:     Launched map tasks=3
08/01/18 19:08:35 INFO mapred.JobClient:     Launched reduce tasks=1
08/01/18 19:08:35 INFO mapred.JobClient:     Data-local map tasks=2
08/01/18 19:08:35 INFO mapred.JobClient:   Map-Reduce Framework
08/01/18 19:08:35 INFO mapred.JobClient:     Map input records=1
08/01/18 19:08:35 INFO mapred.JobClient:     Map output records=1
08/01/18 19:08:35 INFO mapred.JobClient:     Map input bytes=2
08/01/18 19:08:35 INFO mapred.JobClient:     Map output bytes=5
08/01/18 19:08:35 INFO mapred.JobClient:     Reduce input groups=1
08/01/18 19:08:35 INFO mapred.JobClient:     Reduce input records=1
08/01/18 19:08:35 INFO mapred.JobClient:     Reduce output records=1
DataLines       Maps    Reduces AvgTime (milliseconds)
1               2       1       926333



On Fri, 18 Jan 2008, Konstantin Shvachko wrote:

Looks like we still have this unsolved mysterious problem:

http://issues.apache.org/jira/browse/HADOOP-1374

Could it be related to HADOOP-1246? Arun?

Thanks,
--Konstantin

Yunhong Gu1 wrote:

Hi,

If someone knows how to fix the problem described below,
please help
me out. Thanks!

I am testing Hadoop on 2-node cluster and the "reduce"
always hangs
at some stage, even if I use different clusters. My OS is Debian
Linux kernel 2.6 (AMD Opteron w/ 4GB Mem). Hadoop verision
is 0.15.2.
Java version is 1.5.0_01-b08.

I simply tried "./bin/hadoop jar hadoop-0.15.2-test.jar
mrbench" and
when the map stage finishes, the reduce stage will hang
somewhere in
the middle, sometimes at 0%. I also tried any other
mapreduce program
I can find in the example jar package but they all hang.

The log file simply print
2008-01-18 15:15:50,831 INFO org.apache.hadoop.mapred.TaskTracker:
task_200801181424_0004_r_000000_0 0.0% reduce > copy >
2008-01-18 15:15:56,841 INFO org.apache.hadoop.mapred.TaskTracker:
task_200801181424_0004_r_000000_0 0.0% reduce > copy >
2008-01-18 15:16:02,850 INFO org.apache.hadoop.mapred.TaskTracker:
task_200801181424_0004_r_000000_0 0.0% reduce > copy >

forever.

The program does work if I start Hadoop only on single node.

Below is my hadoop-site.xml configuration:

<configuration>

<property>
   <name> fs.default.name</name>
   <value>10.0.0.1:60000</value>
</property>

<property>
   <name>mapred.job.tracker</name>
   <value>10.0.0.1:60001</value>
</property>

<property>
   <name>dfs.data.dir</name>
   <value>/raid/hadoop/data</value>
</property>

<property>
   <name>mapred.local.dir</name>
   <value>/raid/hadoop/mapred</value>
</property>

<property>
  <name>hadoop.tmp.dir</name>
  <value>/raid/hadoop/tmp</value>
</property>

<property>
  <name>mapred.child.java.opts</name>
  <value>-Xmx1024m</value>
</property>

<property>
  <name>mapred.tasktracker.tasks.maximum</name>
  <value>4</value>
</property>

<!--
<property>
  <name>mapred.map.tasks</name>
  <value>7</value>
</property>

<property>
  <name>mapred.reduce.tasks</name>
  <value>3</value>
</property>
-->

<property>
  <name>fs.inmemory.size.mb</name>
  <value>200</value>
</property>

<property>
  <name>dfs.block.size</name>
  <value>134217728</value>
</property>

<property>
  <name>io.sort.factor</name>
  <value>100</value>
</property>

<property>
  <name>io.sort.mb</name>
  <value>200</value>
</property>

<property>
  <name>io.file.buffer.size</name>
  <value>131072</value>
</property>

</configuration>











--
Taeho Kang [tkang.blogspot.com]
Software Engineer, NHN Corporation, Korea

Reply via email to