Thanks Runping. It seems the bug is still open. However, in my case, there were more than plenty of disk space available.
On Jan 16, 2008 2:44 AM, Runping Qi <[EMAIL PROTECTED]> wrote: > > I encountered a similar case. > Here is the Jira: https://issues.apache.org/jira/browse/HADOOP-2164 > > Runping > > > > > -----Original Message----- > > From: Vadim Zaliva [mailto:[EMAIL PROTECTED] > > Sent: Tuesday, January 15, 2008 9:59 PM > > To: hadoop-user@lucene.apache.org > > Subject: Re: unable to figure out this exception from reduce task > > > > On Jan 15, 2008, at 21:53, Jim the Standing Bear wrote: > > > > I was asking lot of questions today, so I am glad to contribute at > > least one answer. I have this problem when there was type mismatch > > for key or values. You need to set up right type at your JobConf like > > this: > > > > conf.setOutputKeyClass(Text.class); > > conf.setOutputValueClass(LongWritable.class); > > > > (using appropriate types our mapper produce) > > > > Vadim > > > > > I am using hadoop 0.15.1 to index some catalog that has a tree-like > > > structure, where the leaf nodes are data files. My main task is a > > > loop that performs a breadth-first walkthrough that parses out URLs > to > > > catalogs and datafiles at the next level, which is done in a mapper. > > > To determine when the loop should terminate, I use a reduce task > that > > > counts the number of new catalogs found, and stops the loop when the > > > count is 0. > > > > > > But while I was running the jobs, I kept getting this exception > > > (pasted below from the logs). I didn't quite understand what it was > > > trying to say. But in my code, I never used LongWritable. Only > Text > > > for output key and output values, and KeyValueTextInputFormat for > > > input. > > > > > > What's weirder is that this exception occurs at different places > from > > > job to job. Sometimes it may be thrown at the 2nd iteration of my > > > loop, while other times, it may be the 3rd, the 4th etc. Can > someone > > > explain to me what and why this is? Also, what would be the best > way > > > to test/debug a hadoop job?? Thanks. > > > > > > > > > 2008-01-16 00:37:19,941 INFO org.apache.hadoop.mapred.ReduceTask: > > > task_200801160024_0011_r_000000_1 Copying > > > task_200801160024_0011_m_000000_0 output from ginkgo.mycluster.org > > > 2008-01-16 00:37:19,953 INFO org.apache.hadoop.mapred.ReduceTask: > > > task_200801160024_0011_r_000000_1 done copying > > > task_200801160024_0011_m_000000_0 output from ginkgo.mycluster.org > > > 2008-01-16 00:37:19,955 INFO org.apache.hadoop.mapred.ReduceTask: > > > task_200801160024_0011_r_000000_1 Copying of all map outputs > complete. > > > Initiating the last merge on the remaining files in > > > ramfs://mapoutput26453615 > > > 2008-01-16 00:37:20,088 WARN org.apache.hadoop.mapred.ReduceTask: > > > task_200801160024_0011_r_000000_1 Final merge of the inmemory files > > > threw an exception: java.io.IOException: java.io.IOException: wrong > > > key class: class org.apache.hadoop.io.LongWritable is not class > > > org.apache.hadoop.io.Text > > > at org.apache.hadoop.io.SequenceFile$Sorter > > > $SegmentDescriptor.nextRawKey(SequenceFile.java:2874) > > > at org.apache.hadoop.io.SequenceFile$Sorter > > > $MergeQueue.merge(SequenceFile.java:2683) > > > at > org.apache.hadoop.io.SequenceFile$Sorter.merge(SequenceFile.java: > > > 2437) > > > at org.apache.hadoop.mapred.ReduceTask > > > $ReduceCopier.fetchOutputs(ReduceTask.java:1153) > > > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:252) > > > at > org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java: > > > 1760) > > > > > > at org.apache.hadoop.mapred.ReduceTask > > > $ReduceCopier.fetchOutputs(ReduceTask.java:1161) > > > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:252) > > > at > org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java: > > > 1760) > > > > > > 2008-01-16 00:37:20,090 WARN org.apache.hadoop.mapred.TaskTracker: > > > Error running child > > > java.io.IOException: task_200801160024_0011_r_000000_1The reduce > > > copier failed > > > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:253) > > > at > org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java: > > > 1760) > > > > > > > > > > > > -- > > > -------------------------------------- > > > Standing Bear Has Spoken > > > -------------------------------------- > > -- -------------------------------------- Standing Bear Has Spoken --------------------------------------