Hi James, If you’re on Spark 1.3 you can use the kill command in spark-submit to shut it down. You’ll need the driver id from the Spark UI or from when you submitted the app.
spark-submit --master spark://master:7077 --kill <driver-id> Thanks, Silvio From: James King Date: Wednesday, May 6, 2015 at 12:02 PM To: user Subject: Stop Cluster Mode Running App I submitted a Spark Application in cluster mode and now every time I stop the cluster and restart it the job resumes execution. I even killed a daemon called DriverWrapper it stops the app but it resumes again. How can stop this application from running?
