You would generally use --conf to set this on the command line if using the shell.
On Tue, Sep 13, 2016, 19:22 Kevin Burton <bur...@spinn3r.com> wrote: > The problem is that without a new spark context, with a custom conf, > elasticsearch-hadoop is refusing to read in settings about the ES setup... > > if I do a sc.stop() , then create a new one, it seems to work fine. > > But it isn't really documented anywhere and all the existing documentation > is now invalid because you get an exception when you try to create a new > spark context. > > On Tue, Sep 13, 2016 at 11:13 AM, Mich Talebzadeh < > mich.talebza...@gmail.com> wrote: > >> I think this works in a shell but you need to allow multiple spark >> contexts >> >> Spark context Web UI available at http://50.140.197.217:55555 >> Spark context available as 'sc' (master = local, app id = >> local-1473789661846). >> Spark session available as 'spark'. >> Welcome to >> ____ __ >> / __/__ ___ _____/ /__ >> _\ \/ _ \/ _ `/ __/ '_/ >> /___/ .__/\_,_/_/ /_/\_\ version 2.0.0 >> /_/ >> Using Scala version 2.11.8 (Java HotSpot(TM) 64-Bit Server VM, Java >> 1.8.0_77) >> Type in expressions to have them evaluated. >> Type :help for more information. >> >> scala> import org.apache.spark.SparkContext >> import org.apache.spark.SparkContext >> scala> val conf = new >> SparkConf().setMaster("local[2]").setAppName("CountingSheep"). >> *set("spark.driver.allowMultipleContexts", "true")*conf: >> org.apache.spark.SparkConf = org.apache.spark.SparkConf@bb5f9d >> scala> val sc = new SparkContext(conf) >> sc: org.apache.spark.SparkContext = org.apache.spark.SparkContext@4888425d >> >> >> HTH >> >> >> Dr Mich Talebzadeh >> >> >> >> LinkedIn * >> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >> >> >> >> http://talebzadehmich.wordpress.com >> >> >> *Disclaimer:* Use it at your own risk. Any and all responsibility for >> any loss, damage or destruction of data or any other property which may >> arise from relying on this email's technical content is explicitly >> disclaimed. The author will in no case be liable for any monetary damages >> arising from such loss, damage or destruction. >> >> >> >> On 13 September 2016 at 18:57, Sean Owen <so...@cloudera.com> wrote: >> >>> But you're in the shell there, which already has a SparkContext for you >>> as sc. >>> >>> On Tue, Sep 13, 2016 at 6:49 PM, Kevin Burton <bur...@spinn3r.com> >>> wrote: >>> >>>> I'm rather confused here as to what to do about creating a new >>>> SparkContext. >>>> >>>> Spark 2.0 prevents it... (exception included below) >>>> >>>> yet a TON of examples I've seen basically tell you to create a new >>>> SparkContext as standard practice: >>>> >>>> >>>> http://spark.apache.org/docs/latest/configuration.html#dynamically-loading-spark-properties >>>> >>>> val conf = new SparkConf() >>>> .setMaster("local[2]") >>>> .setAppName("CountingSheep")val sc = new SparkContext(conf) >>>> >>>> >>>> I'm specifically running into a problem in that ES hadoop won't work >>>> with its settings and I think its related to this problme. >>>> >>>> Do we have to call sc.stop() first and THEN create a new spark context? >>>> >>>> That works,, but I can't find any documentation anywhere telling us the >>>> right course of action. >>>> >>>> >>>> >>>> scala> val sc = new SparkContext(); >>>> org.apache.spark.SparkException: Only one SparkContext may be running >>>> in this JVM (see SPARK-2243). To ignore this error, set >>>> spark.driver.allowMultipleContexts = true. The currently running >>>> SparkContext was created at: >>>> >>>> org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:823) >>>> org.apache.spark.repl.Main$.createSparkSession(Main.scala:95) >>>> <init>(<console>:15) >>>> <init>(<console>:31) >>>> <init>(<console>:33) >>>> .<init>(<console>:37) >>>> .<clinit>(<console>) >>>> .$print$lzycompute(<console>:7) >>>> .$print(<console>:6) >>>> $print(<console>) >>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>> >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> java.lang.reflect.Method.invoke(Method.java:497) >>>> scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:786) >>>> scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1047) >>>> scala.tools.nsc.interpreter.IM >>>> ain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:638) >>>> scala.tools.nsc.interpreter.IM >>>> ain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:637) >>>> scala.reflect.internal.util.Sc >>>> alaClassLoader$class.asContext(ScalaClassLoader.scala:31) >>>> >>>> scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:19) >>>> at >>>> org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$2.apply(SparkContext.scala:2221) >>>> at >>>> org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$2.apply(SparkContext.scala:2217) >>>> at scala.Option.foreach(Option.scala:257) >>>> at >>>> org.apache.spark.SparkContext$.assertNoOtherContextIsRunning(SparkContext.scala:2217) >>>> at >>>> org.apache.spark.SparkContext$.markPartiallyConstructed(SparkContext.scala:2290) >>>> at org.apache.spark.SparkContext.<init>(SparkContext.scala:89) >>>> at org.apache.spark.SparkContext.<init>(SparkContext.scala:121) >>>> ... 48 elided >>>> >>>> >>>> -- >>>> >>>> We’re hiring if you know of any awesome Java Devops or Linux Operations >>>> Engineers! >>>> >>>> Founder/CEO Spinn3r.com >>>> Location: *San Francisco, CA* >>>> blog: http://burtonator.wordpress.com >>>> … or check out my Google+ profile >>>> <https://plus.google.com/102718274791889610666/posts> >>>> >>>> >>> >> > > > -- > > We’re hiring if you know of any awesome Java Devops or Linux Operations > Engineers! > > Founder/CEO Spinn3r.com > Location: *San Francisco, CA* > blog: http://burtonator.wordpress.com > … or check out my Google+ profile > <https://plus.google.com/102718274791889610666/posts> > >