It sounds like you should be writing an application and not trying to force the spark-shell to do more than what it was intended for.
On Tue, Sep 13, 2016 at 11:53 AM, Kevin Burton <bur...@spinn3r.com> wrote: > I sort of agree but the problem is that some of this should be code. > > Some of our ES indexes have 100-200 columns. > > Defining which ones are arrays on the command line is going to get ugly > fast. > > > > On Tue, Sep 13, 2016 at 11:50 AM, Sean Owen <so...@cloudera.com> wrote: > >> You would generally use --conf to set this on the command line if using >> the shell. >> >> >> On Tue, Sep 13, 2016, 19:22 Kevin Burton <bur...@spinn3r.com> wrote: >> >>> The problem is that without a new spark context, with a custom conf, >>> elasticsearch-hadoop is refusing to read in settings about the ES setup... >>> >>> if I do a sc.stop() , then create a new one, it seems to work fine. >>> >>> But it isn't really documented anywhere and all the existing >>> documentation is now invalid because you get an exception when you try to >>> create a new spark context. >>> >>> On Tue, Sep 13, 2016 at 11:13 AM, Mich Talebzadeh < >>> mich.talebza...@gmail.com> wrote: >>> >>>> I think this works in a shell but you need to allow multiple spark >>>> contexts >>>> >>>> Spark context Web UI available at http://50.140.197.217:55555 >>>> Spark context available as 'sc' (master = local, app id = >>>> local-1473789661846). >>>> Spark session available as 'spark'. >>>> Welcome to >>>> ____ __ >>>> / __/__ ___ _____/ /__ >>>> _\ \/ _ \/ _ `/ __/ '_/ >>>> /___/ .__/\_,_/_/ /_/\_\ version 2.0.0 >>>> /_/ >>>> Using Scala version 2.11.8 (Java HotSpot(TM) 64-Bit Server VM, Java >>>> 1.8.0_77) >>>> Type in expressions to have them evaluated. >>>> Type :help for more information. >>>> >>>> scala> import org.apache.spark.SparkContext >>>> import org.apache.spark.SparkContext >>>> scala> val conf = new SparkConf().setMaster("local[2 >>>> ]").setAppName("CountingSheep"). >>>> *set("spark.driver.allowMultipleContexts", "true")*conf: >>>> org.apache.spark.SparkConf = org.apache.spark.SparkConf@bb5f9d >>>> scala> val sc = new SparkContext(conf) >>>> sc: org.apache.spark.SparkContext = org.apache.spark.SparkContext@ >>>> 4888425d >>>> >>>> >>>> HTH >>>> >>>> >>>> Dr Mich Talebzadeh >>>> >>>> >>>> >>>> LinkedIn * >>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >>>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >>>> >>>> >>>> >>>> http://talebzadehmich.wordpress.com >>>> >>>> >>>> *Disclaimer:* Use it at your own risk. Any and all responsibility for >>>> any loss, damage or destruction of data or any other property which may >>>> arise from relying on this email's technical content is explicitly >>>> disclaimed. The author will in no case be liable for any monetary damages >>>> arising from such loss, damage or destruction. >>>> >>>> >>>> >>>> On 13 September 2016 at 18:57, Sean Owen <so...@cloudera.com> wrote: >>>> >>>>> But you're in the shell there, which already has a SparkContext for >>>>> you as sc. >>>>> >>>>> On Tue, Sep 13, 2016 at 6:49 PM, Kevin Burton <bur...@spinn3r.com> >>>>> wrote: >>>>> >>>>>> I'm rather confused here as to what to do about creating a new >>>>>> SparkContext. >>>>>> >>>>>> Spark 2.0 prevents it... (exception included below) >>>>>> >>>>>> yet a TON of examples I've seen basically tell you to create a new >>>>>> SparkContext as standard practice: >>>>>> >>>>>> http://spark.apache.org/docs/latest/configuration.html#dynam >>>>>> ically-loading-spark-properties >>>>>> >>>>>> val conf = new SparkConf() >>>>>> .setMaster("local[2]") >>>>>> .setAppName("CountingSheep")val sc = new SparkContext(conf) >>>>>> >>>>>> >>>>>> I'm specifically running into a problem in that ES hadoop won't work >>>>>> with its settings and I think its related to this problme. >>>>>> >>>>>> Do we have to call sc.stop() first and THEN create a new spark >>>>>> context? >>>>>> >>>>>> That works,, but I can't find any documentation anywhere telling us >>>>>> the right course of action. >>>>>> >>>>>> >>>>>> >>>>>> scala> val sc = new SparkContext(); >>>>>> org.apache.spark.SparkException: Only one SparkContext may be >>>>>> running in this JVM (see SPARK-2243). To ignore this error, set >>>>>> spark.driver.allowMultipleContexts = true. The currently running >>>>>> SparkContext was created at: >>>>>> org.apache.spark.sql.SparkSession$Builder.getOrCreate( >>>>>> SparkSession.scala:823) >>>>>> org.apache.spark.repl.Main$.createSparkSession(Main.scala:95) >>>>>> <init>(<console>:15) >>>>>> <init>(<console>:31) >>>>>> <init>(<console>:33) >>>>>> .<init>(<console>:37) >>>>>> .<clinit>(<console>) >>>>>> .$print$lzycompute(<console>:7) >>>>>> .$print(<console>:6) >>>>>> $print(<console>) >>>>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce >>>>>> ssorImpl.java:62) >>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe >>>>>> thodAccessorImpl.java:43) >>>>>> java.lang.reflect.Method.invoke(Method.java:497) >>>>>> scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:786) >>>>>> scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.s >>>>>> cala:1047) >>>>>> scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$lo >>>>>> adAndRunReq$1.apply(IMain.scala:638) >>>>>> scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$lo >>>>>> adAndRunReq$1.apply(IMain.scala:637) >>>>>> scala.reflect.internal.util.ScalaClassLoader$class.asContext >>>>>> (ScalaClassLoader.scala:31) >>>>>> scala.reflect.internal.util.AbstractFileClassLoader.asContex >>>>>> t(AbstractFileClassLoader.scala:19) >>>>>> at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextI >>>>>> sRunning$2.apply(SparkContext.scala:2221) >>>>>> at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextI >>>>>> sRunning$2.apply(SparkContext.scala:2217) >>>>>> at scala.Option.foreach(Option.scala:257) >>>>>> at org.apache.spark.SparkContext$.assertNoOtherContextIsRunning >>>>>> (SparkContext.scala:2217) >>>>>> at org.apache.spark.SparkContext$.markPartiallyConstructed(Spar >>>>>> kContext.scala:2290) >>>>>> at org.apache.spark.SparkContext.<init>(SparkContext.scala:89) >>>>>> at org.apache.spark.SparkContext.<init>(SparkContext.scala:121) >>>>>> ... 48 elided >>>>>> >>>>>> >>>>>> -- >>>>>> >>>>>> We’re hiring if you know of any awesome Java Devops or Linux >>>>>> Operations Engineers! >>>>>> >>>>>> Founder/CEO Spinn3r.com >>>>>> Location: *San Francisco, CA* >>>>>> blog: http://burtonator.wordpress.com >>>>>> … or check out my Google+ profile >>>>>> <https://plus.google.com/102718274791889610666/posts> >>>>>> >>>>>> >>>>> >>>> >>> >>> >>> -- >>> >>> We’re hiring if you know of any awesome Java Devops or Linux Operations >>> Engineers! >>> >>> Founder/CEO Spinn3r.com >>> Location: *San Francisco, CA* >>> blog: http://burtonator.wordpress.com >>> … or check out my Google+ profile >>> <https://plus.google.com/102718274791889610666/posts> >>> >>> > > > -- > > We’re hiring if you know of any awesome Java Devops or Linux Operations > Engineers! > > Founder/CEO Spinn3r.com > Location: *San Francisco, CA* > blog: http://burtonator.wordpress.com > … or check out my Google+ profile > <https://plus.google.com/102718274791889610666/posts> > >