Hi Duc, Are you running Spark on Mesos with cluster mode? And what's your cluster mode submission, and version of Spark are you running?
Tim On Sat, Jan 30, 2016 at 8:19 AM, PhuDuc Nguyen <duc.was.h...@gmail.com> wrote: > I have a spark job running on Mesos in multi-master and supervise mode. If > I kill it, it is resilient as expected and respawns on another node. > However, I cannot kill it when I need to. I have tried 2 methods: > > 1) ./bin/spark-class org.apache.spark.deploy.Client kill > <masterIp:port???> <driverID> > > 2) ./bin/spark-submit --master mesos://<masterIp:7077> --kill <driverID> > > Method 2, accepts the kill request but is respawned on another node. > Ultimately, I can't get either method to kill the job. I suspect I have > the wrong port for the master URL during the kill request for method 1? > I've tried every combination of IP and port I can think of, is there one I > am missing? > > Ports I've tried: > 5050 = mesos UI > 8080 = marathon > 7077 = spark dispatcher > 8081 = spark drivers UI > 4040 = spark job UI > > thanks, > Duc >