I remember there was some issue with the above command in previous veresions of spark. Its nice that its working now :)
On Wed, Nov 12, 2014 at 5:50 PM, Tao Xiao <xiaotao.cs....@gmail.com> wrote: > Thanks for your replies. > > Actually we can kill a driver by the command "bin/spark-class > org.apache.spark.deploy.Client kill <spark-master> <driver-id>" if you > know the driver id. > > 2014-11-11 22:35 GMT+08:00 Ritesh Kumar Singh < > riteshoneinamill...@gmail.com>: > >> There is a property : >> spark.ui.killEnabled >> which needs to be set true for killing applications directly from the >> webUI. >> Check the link: >> Kill Enable spark job >> <http://spark.apache.org/docs/latest/configuration.html#spark-ui> >> >> Thanks >> >> On Tue, Nov 11, 2014 at 7:42 PM, Sonal Goyal <sonalgoy...@gmail.com> >> wrote: >> >>> The web interface has a kill link. You can try using that. >>> >>> Best Regards, >>> Sonal >>> Founder, Nube Technologies <http://www.nubetech.co> >>> >>> <http://in.linkedin.com/in/sonalgoyal> >>> >>> >>> >>> On Tue, Nov 11, 2014 at 7:28 PM, Tao Xiao <xiaotao.cs....@gmail.com> >>> wrote: >>> >>>> I'm using Spark 1.0.0 and I'd like to kill a job running in cluster >>>> mode, which means the driver is not running on local node. >>>> >>>> So how can I kill such a job? Is there a command like "hadoop job >>>> -kill <job-id>" which kills a running MapReduce job ? >>>> >>>> Thanks >>>> >>> >>> >> >