[
https://issues.apache.org/jira/browse/HADOOP-1193?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Arun C Murthy updated HADOOP-1193:
----------------------------------
Attachment: HADOOP-1193_2_20070524.patch
Here is an updated version of the patch with the changes I made to BigMapOutput
to help test it (basically made it extend ToolBase and added an option to
create the large map input too)...
I have tested this with large map inputs (>2G) and seems to hold up well i.e.
the codec pool ensure we create only 1 compressor and very small no. of
decompressors (less than 10) even for extremely large map inputs (>2G).
> Map/reduce job gets OutOfMemoryException when set map out to be compressed
> --------------------------------------------------------------------------
>
> Key: HADOOP-1193
> URL: https://issues.apache.org/jira/browse/HADOOP-1193
> Project: Hadoop
> Issue Type: Bug
> Components: mapred
> Affects Versions: 0.12.2
> Reporter: Hairong Kuang
> Assigned To: Arun C Murthy
> Attachments: HADOOP-1193_1_20070517.patch,
> HADOOP-1193_2_20070524.patch
>
>
> One of my jobs quickly fails with the OutOfMemoryException when I set the map
> out to be compressed. But it worked fine with release 0.10.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.