Hi Duc,

Are you running Spark on Mesos with cluster mode? And what's your cluster
mode submission, and version of Spark are you running?

Tim

On Sat, Jan 30, 2016 at 8:19 AM, PhuDuc Nguyen <duc.was.h...@gmail.com>
wrote:

> I have a spark job running on Mesos in multi-master and supervise mode. If
> I kill it, it is resilient as expected and respawns on another node.
> However, I cannot kill it when I need to. I have tried 2 methods:
>
> 1) ./bin/spark-class org.apache.spark.deploy.Client kill
> <masterIp:port???> <driverID>
>
> 2) ./bin/spark-submit --master mesos://<masterIp:7077> --kill <driverID>
>
> Method 2, accepts the kill request but is respawned on another node.
> Ultimately, I can't get either method to kill the job. I suspect I have
> the wrong port for the master URL during the kill request for method 1?
> I've tried every combination of IP and port I can think of, is there one I
> am missing?
>
> Ports I've tried:
> 5050 = mesos UI
> 8080 = marathon
> 7077 = spark dispatcher
> 8081 = spark drivers UI
> 4040 = spark job UI
>
> thanks,
> Duc
>

Reply via email to