You can create a custom mesos framework for your requirement, to get you started you can check this out http://mesos.apache.org/documentation/latest/app-framework-development-guide/
Thanks Best Regards On Mon, Aug 24, 2015 at 12:11 PM, Romi Kuntsman <r...@totango.com> wrote: > Hi, > I have a spark standalone cluster with 100s of applications per day, and > it changes size (more or less workers) at various hours. The driver runs on > a separate machine outside the spark cluster. > > When a job is running and it's worker is killed (because at that hour the > number of workers is reduced), it sometimes fails, instead of > redistributing the work to other workers. > > How is it possible to decomission a worker, so that it doesn't receive any > new work, but does finish all existing work before shutting down? > > Thanks! >