Hi all, when runnig the Spark cluster in standalone mode I am able to create the Spark context from Java via the following code snippet:
SparkConf conf = new SparkConf() > .setAppName("MySparkApp") > .setMaster("spark://SPARK_MASTER:7077") > .setJars(jars); > JavaSparkContext sc = new JavaSparkContext(conf); As soon as I'm done with my processing, I can just close it via > sc.stop(); > Now my question: Is the same also possible when running Spark on YARN? I currently don't see how this should be possible without submitting your application as a packaged jar file. Is there a way to get this kind of interactivity from within your Scala/Java code? Regards, Andrea