[ https://issues.apache.org/jira/browse/MAPREDUCE-5862?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13983935#comment-13983935 ]
Sandy Ryza commented on MAPREDUCE-5862: --------------------------------------- {code} + checkRecordSpanningMultipleSplits("recordSpanningMultipleSplits.txt.bz2", + 200 * 1000, + true); {code} indentation should be: {code} + checkRecordSpanningMultipleSplits("recordSpanningMultipleSplits.txt.bz2", + 200 * 1000, true); {code} I can fix these on commit. Otherwise, the updated patch looks good to me. [~jlowe], anything you see that I'm missing? > Line records longer than 2x split size aren't handled correctly > --------------------------------------------------------------- > > Key: MAPREDUCE-5862 > URL: https://issues.apache.org/jira/browse/MAPREDUCE-5862 > Project: Hadoop Map/Reduce > Issue Type: Bug > Affects Versions: 2.3.0 > Reporter: bc Wong > Assignee: bc Wong > Priority: Critical > Attachments: 0001-Handle-records-larger-than-2x-split-size.1.patch, > 0001-Handle-records-larger-than-2x-split-size.patch, > 0001-Handle-records-larger-than-2x-split-size.patch, > recordSpanningMultipleSplits.txt.bz2 > > > Suppose this split (100-200) is in the middle of a record (90-240): > {noformat} > 0 100 200 300 > |---- split ----|---- curr ----|---- split ----| > <------- record -------> > 90 240 > {noformat} > > Currently, the first split would read the entire record, up to offset 240, > which is good. But the 2nd split has a bug in producing a phantom record of > (200, 240). -- This message was sent by Atlassian JIRA (v6.2#6252)