Sounds like you are in Yarn-Cluster mode. I created a JIRA SPARK-3913 <https://issues.apache.org/jira/browse/SPARK-3913> and PR https://github.com/apache/spark/pull/2786
is this what you looking for ? Chester On Sat, May 2, 2015 at 10:32 PM, Yijie Shen <henry.yijies...@gmail.com> wrote: > Hi, > > I’ve posted this problem in user@spark but find no reply, therefore moved > to dev@spark, sorry for duplication. > > I am wondering if it is possible to submit, monitor & kill spark > applications from another service. > > I have wrote a service this: > > parse user commands > translate them into understandable arguments to an already prepared > Spark-SQL application > submit the application along with arguments to Spark Cluster > using spark-submit from ProcessBuilder > run generated applications' driver in cluster mode. > The above 4 steps has been finished, but I have difficulties in these two: > > Query about the applications status, for example, the percentage > completion. > Kill queries accordingly > What I find in spark standalone documentation suggest kill application > using: > > ./bin/spark-class org.apache.spark.deploy.Client kill <master url> <driver > ID> > > And should find > the driver ID through the standalone Master web UI at > http://<master url>:8080. > > Are there any programmatically methods I could get the driverID submitted > by my `ProcessBuilder` and query status about the query? > > Any Suggestions? > > — > Best Regards! > Yijie Shen