[
https://issues.apache.org/jira/browse/PIG-3480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13781241#comment-13781241
]
Rohini Palaniswamy commented on PIG-3480:
-----------------------------------------
[~aniket486],
Would prefer having TFile as the default and Sequence file as an option. We
have not had any issues with it for years and may be few others too it has been
the default. Also the performance numbers are better for it by 10-40% with
compression according to HADOOP-3315 that you have referred.
And it would also be good to have the actual cause of failure with TFile
investigated if possible to see if something is not being done right in Pig as
TFile just writes byte[] for keys and values.
> TFile-based tmpfile compression crashes in some cases
> -----------------------------------------------------
>
> Key: PIG-3480
> URL: https://issues.apache.org/jira/browse/PIG-3480
> Project: Pig
> Issue Type: Bug
> Reporter: Dmitriy V. Ryaboy
> Fix For: 0.12.0
>
> Attachments: PIG-3480.patch
>
>
> When pig tmpfile compression is on, some jobs fail inside core hadoop
> internals.
> Suspect TFile is the problem, because an experiment in replacing TFile with
> SequenceFile succeeded.
--
This message was sent by Atlassian JIRA
(v6.1#6144)