[ https://issues.apache.org/jira/browse/SPARK-32571?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Dongjoon Hyun updated SPARK-32571: ---------------------------------- Priority: Major (was: Critical) > yarnClient.killApplication(appId) is never called > ------------------------------------------------- > > Key: SPARK-32571 > URL: https://issues.apache.org/jira/browse/SPARK-32571 > Project: Spark > Issue Type: Bug > Components: Spark Submit, YARN > Affects Versions: 2.4.0, 3.0.0 > Reporter: A Tester > Priority: Major > > *Problem Statement:* > When an application is submitted using spark-submit in cluster mode using > yarn, the spark application continues to run on the cluster, even if > spark-submit itself has been requested to shutdown (Ctrl-C/SIGTERM/etc.) > While there is code inside org.apache.spark.deploy.yarn.Client.scala that > would lead you to believe the spark application on the cluster will shut > down, this code is not currently reachable. > Example of behavior: > spark-submit ... > <Ctrl-C> or kill -15 <pid> > spark-submit itself dies > job can still be found running on the cluster > > *Expectation:* > When spark-submit is in monitoring a yarn app and spark-submit itself is > requested to shutdown (SIGTERM, HUP,etc.), it should call > yarnClient.killApplication(appId) so that the actual spark application > running on the cluster is killed. > > > *Proposal* > There is already a shutdown hook registered which cleans up temp files. > Could this be extended to call yarnClient.killApplication? > I believe the default behavior should be to request yarn to kill the > application, however I can imagine use cases where you may still want it to > run. So facilitate these use cases, an option should be provided to skip > this hook. > -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org