[ 
https://issues.apache.org/jira/browse/SPARK-20045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hyukjin Kwon updated SPARK-20045:
---------------------------------
    Labels: bulk-closed  (was: )

> Make sure SparkHadoopMapReduceWriter is resilient to failures of writers and 
> committers
> ---------------------------------------------------------------------------------------
>
>                 Key: SPARK-20045
>                 URL: https://issues.apache.org/jira/browse/SPARK-20045
>             Project: Spark
>          Issue Type: Improvement
>          Components: Spark Core
>    Affects Versions: 2.2.0
>            Reporter: Steve Loughran
>            Priority: Minor
>              Labels: bulk-closed
>
> Following on from SPARK-20038: review {{SparkHadoopMapReduceWriter}} and 
> ensure that it's failure handling code is itself resilient to follow on 
> failures, especially in things like {{writer.close()}} and the 
> abortTask/abortJob calls. That's to ensure as robust a cleanup as possible, 
> and to stop the original exception getting lost.
> At a quick glance 
> 1. {{executeTask()}}'s catch logic should catch & log any failure in the 
> {{writer.close()}} 
> 2. The Hadoop commit protocol's abort* operation's can throw IOEs. Again, 
> they need to be caught and logged
> Should be testable with mocking, and worthwhile giving how important commit 
> protocol resilience is.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to