You would generally use --conf to set this on the command line if using the
shell.

On Tue, Sep 13, 2016, 19:22 Kevin Burton <bur...@spinn3r.com> wrote:

> The problem is that without a new spark context, with a custom conf,
> elasticsearch-hadoop is refusing to read in settings about the ES setup...
>
> if I do a sc.stop() , then create a new one, it seems to work fine.
>
> But it isn't really documented anywhere and all the existing documentation
> is now invalid because you get an exception when you try to create a new
> spark context.
>
> On Tue, Sep 13, 2016 at 11:13 AM, Mich Talebzadeh <
> mich.talebza...@gmail.com> wrote:
>
>> I think this works in a shell but you need to allow multiple spark
>> contexts
>>
>> Spark context Web UI available at http://50.140.197.217:55555
>> Spark context available as 'sc' (master = local, app id =
>> local-1473789661846).
>> Spark session available as 'spark'.
>> Welcome to
>>       ____              __
>>      / __/__  ___ _____/ /__
>>     _\ \/ _ \/ _ `/ __/  '_/
>>    /___/ .__/\_,_/_/ /_/\_\   version 2.0.0
>>       /_/
>> Using Scala version 2.11.8 (Java HotSpot(TM) 64-Bit Server VM, Java
>> 1.8.0_77)
>> Type in expressions to have them evaluated.
>> Type :help for more information.
>>
>> scala> import org.apache.spark.SparkContext
>> import org.apache.spark.SparkContext
>> scala>  val conf = new
>> SparkConf().setMaster("local[2]").setAppName("CountingSheep").
>> *set("spark.driver.allowMultipleContexts", "true")*conf:
>> org.apache.spark.SparkConf = org.apache.spark.SparkConf@bb5f9d
>> scala> val sc = new SparkContext(conf)
>> sc: org.apache.spark.SparkContext = org.apache.spark.SparkContext@4888425d
>>
>>
>> HTH
>>
>>
>> Dr Mich Talebzadeh
>>
>>
>>
>> LinkedIn * 
>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>>
>>
>>
>> http://talebzadehmich.wordpress.com
>>
>>
>> *Disclaimer:* Use it at your own risk. Any and all responsibility for
>> any loss, damage or destruction of data or any other property which may
>> arise from relying on this email's technical content is explicitly
>> disclaimed. The author will in no case be liable for any monetary damages
>> arising from such loss, damage or destruction.
>>
>>
>>
>> On 13 September 2016 at 18:57, Sean Owen <so...@cloudera.com> wrote:
>>
>>> But you're in the shell there, which already has a SparkContext for you
>>> as sc.
>>>
>>> On Tue, Sep 13, 2016 at 6:49 PM, Kevin Burton <bur...@spinn3r.com>
>>> wrote:
>>>
>>>> I'm rather confused here as to what to do about creating a new
>>>> SparkContext.
>>>>
>>>> Spark 2.0 prevents it... (exception included below)
>>>>
>>>> yet a TON of examples I've seen basically tell you to create a new
>>>> SparkContext as standard practice:
>>>>
>>>>
>>>> http://spark.apache.org/docs/latest/configuration.html#dynamically-loading-spark-properties
>>>>
>>>> val conf = new SparkConf()
>>>>              .setMaster("local[2]")
>>>>              .setAppName("CountingSheep")val sc = new SparkContext(conf)
>>>>
>>>>
>>>> I'm specifically running into a problem in that ES hadoop won't work
>>>> with its settings and I think its related to this problme.
>>>>
>>>> Do we have to call sc.stop() first and THEN create a new spark context?
>>>>
>>>> That works,, but I can't find any documentation anywhere telling us the
>>>> right course of action.
>>>>
>>>>
>>>>
>>>> scala> val sc = new SparkContext();
>>>> org.apache.spark.SparkException: Only one SparkContext may be running
>>>> in this JVM (see SPARK-2243). To ignore this error, set
>>>> spark.driver.allowMultipleContexts = true. The currently running
>>>> SparkContext was created at:
>>>>
>>>> org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:823)
>>>> org.apache.spark.repl.Main$.createSparkSession(Main.scala:95)
>>>> <init>(<console>:15)
>>>> <init>(<console>:31)
>>>> <init>(<console>:33)
>>>> .<init>(<console>:37)
>>>> .<clinit>(<console>)
>>>> .$print$lzycompute(<console>:7)
>>>> .$print(<console>:6)
>>>> $print(<console>)
>>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>
>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>>
>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>> java.lang.reflect.Method.invoke(Method.java:497)
>>>> scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:786)
>>>> scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1047)
>>>> scala.tools.nsc.interpreter.IM
>>>> ain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:638)
>>>> scala.tools.nsc.interpreter.IM
>>>> ain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:637)
>>>> scala.reflect.internal.util.Sc
>>>> alaClassLoader$class.asContext(ScalaClassLoader.scala:31)
>>>>
>>>> scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:19)
>>>>   at
>>>> org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$2.apply(SparkContext.scala:2221)
>>>>   at
>>>> org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$2.apply(SparkContext.scala:2217)
>>>>   at scala.Option.foreach(Option.scala:257)
>>>>   at
>>>> org.apache.spark.SparkContext$.assertNoOtherContextIsRunning(SparkContext.scala:2217)
>>>>   at
>>>> org.apache.spark.SparkContext$.markPartiallyConstructed(SparkContext.scala:2290)
>>>>   at org.apache.spark.SparkContext.<init>(SparkContext.scala:89)
>>>>   at org.apache.spark.SparkContext.<init>(SparkContext.scala:121)
>>>>   ... 48 elided
>>>>
>>>>
>>>> --
>>>>
>>>> We’re hiring if you know of any awesome Java Devops or Linux Operations
>>>> Engineers!
>>>>
>>>> Founder/CEO Spinn3r.com
>>>> Location: *San Francisco, CA*
>>>> blog: http://burtonator.wordpress.com
>>>> … or check out my Google+ profile
>>>> <https://plus.google.com/102718274791889610666/posts>
>>>>
>>>>
>>>
>>
>
>
> --
>
> We’re hiring if you know of any awesome Java Devops or Linux Operations
> Engineers!
>
> Founder/CEO Spinn3r.com
> Location: *San Francisco, CA*
> blog: http://burtonator.wordpress.com
> … or check out my Google+ profile
> <https://plus.google.com/102718274791889610666/posts>
>
>

Reply via email to