Thanks for your replies.
Actually we can kill a driver by the command bin/spark-class
org.apache.spark.deploy.Client kill spark-master driver-id if you know
the driver id.
2014-11-11 22:35 GMT+08:00 Ritesh Kumar Singh riteshoneinamill...@gmail.com
:
There is a property :
spark.ui.killEnabled
which needs to be set true for killing applications directly from the
webUI.
Check the link:
Kill Enable spark job
http://spark.apache.org/docs/latest/configuration.html#spark-ui
Thanks
On Tue, Nov 11, 2014 at 7:42 PM, Sonal Goyal sonalgoy...@gmail.com
wrote:
The web interface has a kill link. You can try using that.
Best Regards,
Sonal
Founder, Nube Technologies http://www.nubetech.co
http://in.linkedin.com/in/sonalgoyal
On Tue, Nov 11, 2014 at 7:28 PM, Tao Xiao xiaotao.cs@gmail.com
wrote:
I'm using Spark 1.0.0 and I'd like to kill a job running in cluster
mode, which means the driver is not running on local node.
So how can I kill such a job? Is there a command like hadoop job -kill
job-id which kills a running MapReduce job ?
Thanks