I am sure that the hadoop user is identical with me :0.
my job took too long time to complete it.
this is the problem what I have now.
so many failure on a reduce step.
hadoop want to complete it anyway.
go to retry
and failed
go to retry
and failed
...
...
finally, I need to wait about 1 or 2 hours to see "SUCCESS" of my job.
thanks james.
Junyoung Kim (juneng...@gmail.com)
On 05/03/2011 10:56 AM, James Seigel wrote:
Is mapreduce running as the hadoop user? If so it can’t erase the files in
tmp. Which might be causing you some hilarity
:)
J
On 2011-05-02, at 7:43 PM, Jun Young Kim wrote:
Hi,
To James.
this is the permission of hadoop.tmp.dir.
$> ls -al
drwxr-xr-x 6 juneng juneng 4096 5월 3 10:37 hadoop-juneng.2
To Harsh.
yes, out cluster has 96 occupied reducer slots.
and my job is using 90 reduce tasks at one time to complete it.
thanks for all.
Junyoung Kim (juneng...@gmail.com)
On 05/02/2011 08:32 PM, James Seigel wrote:
What are your permissions on your hadoop.tmp.dir ?
James
Sent from my mobile. Please excuse the typos.
On 2011-05-02, at 1:26 AM, Jun Young Kim<juneng...@gmail.com> wrote:
hi, all.
I got so many failures on a reducing step.
see this error.
java.io.IOException: Failed to delete earlier output of task:
attempt_201105021341_0021_r_000001_0
at
org.apache.hadoop.mapred.FileOutputCommitter.moveTaskOutputs(FileOutputCommitter.java:157)
at
org.apache.hadoop.mapred.FileOutputCommitter.moveTaskOutputs(FileOutputCommitter.java:173)
at
org.apache.hadoop.mapred.FileOutputCommitter.moveTaskOutputs(FileOutputCommitter.java:173)
at
org.apache.hadoop.mapred.FileOutputCommitter.commitTask(FileOutputCommitter.java:133)
at
org.apache.hadoop.mapred.OutputCommitter.commitTask(OutputCommitter.java:233)
at org.apache.hadoop.mapred.Task.commit(Task.java:962)
at org.apache.hadoop.mapred.Task.done(Task.java:824)
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:391)
at org.apache.hadoop.mapred.Child$4.run(Child.java:217)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:396)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:742)
at org.apache.hadoop.mapred.C
this error was happened after adopting MultipleTextOutputFormat class in my job.
the job is producing thousands of different output files on a HDFS.
anybody can guess reasons?
thanks.
--
Junyoung Kim (juneng...@gmail.com)