Marcelo Vanzin created SPARK-20904:
--------------------------------------

             Summary: Task failures during shutdown cause problems with 
preempted executors
                 Key: SPARK-20904
                 URL: https://issues.apache.org/jira/browse/SPARK-20904
             Project: Spark
          Issue Type: Bug
          Components: Spark Core, YARN
    Affects Versions: 1.6.0
            Reporter: Marcelo Vanzin


Spark runs tasks in a thread pool that uses daemon threads in each executor. 
That means that when the JVM gets a signal to shut down, those tasks keep 
running.

Now when YARN preempts an executor, it sends a SIGTERM to the process, 
triggering the JVM shutdown. That causes shutdown hooks to run which may cause 
user code running in those tasks to fail, and report task failures to the 
driver. Those failures are then counted towards the maximum number of allowed 
failures, even though in this case we don't want that because the executor was 
preempted.

So we need a better way to handle that situation.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to