[ https://issues.apache.org/jira/browse/MAPREDUCE-5472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Vinod Kumar Vavilapalli moved YARN-1086 to MAPREDUCE-5472: ---------------------------------------------------------- Target Version/s: 2.1.1-beta (was: 2.1.1-beta) Key: MAPREDUCE-5472 (was: YARN-1086) Project: Hadoop Map/Reduce (was: Hadoop YARN) > reducer of sort job restarts from scratch in between after RM restart > --------------------------------------------------------------------- > > Key: MAPREDUCE-5472 > URL: https://issues.apache.org/jira/browse/MAPREDUCE-5472 > Project: Hadoop Map/Reduce > Issue Type: Bug > Reporter: yeshavora > Priority: Blocker > > Steps Followed: > 1) Run a sort job. As soon as it finishes all the map tasks. [100% map], > restart resource manager. > 2) Analyse the progress of the sort job. > It starts with 100% map 0% reduce > 100% map 32% reduce > 100% map 0% reduce > Reducer stays at 30% reduce for around 5-10 minutes. and again start reducer > from scratch. > Log from failed reducer attempt: > Error: java.io.IOException: Error while reading compressed data at > org.apache.hadoop.io.IOUtils.wrappedReadForCompressedData(IOUtils.java:174) > at org.apache.hadoop.mapred.IFile$Reader.readData(IFile.java:383) at > org.apache.hadoop.mapred.IFile$Reader.nextRawValue(IFile.java:444) at > org.apache.hadoop.mapred.Merger$Segment.nextRawValue(Merger.java:327) at > org.apache.hadoop.mapred.Merger$Segment.getValue(Merger.java:309) at > org.apache.hadoop.mapred.Merger$MergeQueue.next(Merger.java:533) at > org.apache.hadoop.mapred.ReduceTask$4.next(ReduceTask.java:619) at > org.apache.hadoop.mapreduce.task.ReduceContextImpl.nextKeyValue(ReduceContextImpl.java:154) > at > org.apache.hadoop.mapreduce.task.ReduceContextImpl.nextKey(ReduceContextImpl.java:121) > at > org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer$Context.nextKey(WrappedReducer.java:297) > at org.apache.hadoop.mapreduce.Reducer.run(Reducer.java:170) at > org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:645) at > org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:405) at > org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:162) at > java.security.AccessController.doPrivileged(Native Method) at > javax.security.auth.Subject.doAs(Subject.java:396) at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1477) > at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:157) Caused by: > org.apache.hadoop.fs.FSError: java.io.IOException: Input/output error at > org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileInputStream.read(RawLocalFileSystem.java:177) > at java.io.BufferedInputStream.read1(BufferedInputStream.java:256) at > java.io.BufferedInputStream.read(BufferedInputStream.java:317) at > java.io.DataInputStream.read(DataInputStream.java:132) at > org.apache.hadoop.mapred.IFileInputStream.doRead(IFileInputStream.java:209) > at org.apache.hadoop.mapred.IFileInputStream.read(IFileInputStream.java:152) > at > org.apache.hadoop.io.compress.BlockDecompressorStream.getCompressedData(BlockDecompressorStream.java:127) > at > org.apache.hadoop.io.compress.BlockDecompressorStream.decompress(BlockDecompressorStream.java:98) > at > org.apache.hadoop.io.compress.DecompressorStream.read(DecompressorStream.java:85) > at > org.apache.hadoop.io.IOUtils.wrappedReadForCompressedData(IOUtils.java:170) > ... 17 more Caused by: java.io.IOException: Input/output error at > java.io.FileInputStream.readBytes(Native Method) at > java.io.FileInputStream.read(FileInputStream.java:220) at > org.apache.hadoop.fs.RawLocalFileSystem$TrackingFileInputStream.read(RawLocalFileSystem.java:110) > at > org.apache.hadoop.fs.RawLocalFileSystem$LocalFSFileInputStream.read(RawLocalFileSystem.java:171) > ... 26 more -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira