[
https://issues.apache.org/jira/browse/SPARK-20045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Hyukjin Kwon updated SPARK-20045:
---------------------------------
Labels: bulk-closed (was: )
> Make sure SparkHadoopMapReduceWriter is resilient to failures of writers and
> committers
> ---------------------------------------------------------------------------------------
>
> Key: SPARK-20045
> URL: https://issues.apache.org/jira/browse/SPARK-20045
> Project: Spark
> Issue Type: Improvement
> Components: Spark Core
> Affects Versions: 2.2.0
> Reporter: Steve Loughran
> Priority: Minor
> Labels: bulk-closed
>
> Following on from SPARK-20038: review {{SparkHadoopMapReduceWriter}} and
> ensure that it's failure handling code is itself resilient to follow on
> failures, especially in things like {{writer.close()}} and the
> abortTask/abortJob calls. That's to ensure as robust a cleanup as possible,
> and to stop the original exception getting lost.
> At a quick glance
> 1. {{executeTask()}}'s catch logic should catch & log any failure in the
> {{writer.close()}}
> 2. The Hadoop commit protocol's abort* operation's can throw IOEs. Again,
> they need to be caught and logged
> Should be testable with mocking, and worthwhile giving how important commit
> protocol resilience is.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]