[ https://issues.apache.org/jira/browse/HADOOP-1837?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Robert Chansler updated HADOOP-1837: ------------------------------------ Component/s: dfs > Insufficient space exception from InMemoryFileSystem after raising > fs.inmemory.size.mb > -------------------------------------------------------------------------------------- > > Key: HADOOP-1837 > URL: https://issues.apache.org/jira/browse/HADOOP-1837 > Project: Hadoop > Issue Type: Bug > Components: dfs > Affects Versions: 0.13.1 > Reporter: Joydeep Sen Sarma > Priority: Minor > > trying out larger in-memory file system (curious if that helped speed the > sort phase). in this run - i had sized it to 500MB. There's plenty of RAM in > the machine (8GB) and the tasks are launched with -Xmx2048 option (so there's > plenty of heap space as well). However - observing this exception: > 2007-09-04 13:47:51,718 INFO org.apache.hadoop.mapred.ReduceTask: > task_0002_r_000002_0 Copying task_0002_m_000124_0 output from hadoop004.sf > 2p.facebook.com. > 2007-09-04 13:47:52,188 WARN org.apache.hadoop.mapred.ReduceTask: > task_0002_r_000002_0 copy failed: task_0002_m_000124_0 from hadoop004.sf2p > .facebook.com > 2007-09-04 13:47:52,189 WARN org.apache.hadoop.mapred.ReduceTask: > java.io.IOException: Insufficient space > at > org.apache.hadoop.fs.InMemoryFileSystem$RawInMemoryFileSystem$InMemoryOutputStream.write(InMemoryFileSystem.java:181) > at > org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:38) > at > java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:65) > at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:123) > at java.io.DataOutputStream.flush(DataOutputStream.java:106) > at > org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:91) > at > org.apache.hadoop.fs.ChecksumFileSystem$FSOutputSummer.close(ChecksumFileSystem.java:416) > at > org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:48) > at > org.apache.hadoop.fs.FSDataOutputStream$Buffer.close(FSDataOutputStream.java:72) > at > org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:92) > at > org.apache.hadoop.mapred.MapOutputLocation.getFile(MapOutputLocation.java:251) > at > org.apache.hadoop.mapred.ReduceTask$ReduceCopier$MapOutputCopier.copyOutput(ReduceTask.java:680) > at > org.apache.hadoop.mapred.ReduceTask$ReduceCopier$MapOutputCopier.run(ReduceTask.java:641) > 2007-09-04 13:47:52,189 WARN org.apache.hadoop.mapred.ReduceTask: > task_0002_r_000002_0 adding host hadoop004.sf2p.facebook.com to penalty bo > x, next contact in 64 seconds > so this ends up slowing stuff down since we backoff on the source host (even > though it's not it's fault). Looking at the code, seems like ReduceTask is > trying to write more to InMemoryFileSystem than it should. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.