Hi, You should have all the deps being "provided" since they're provided by spark infra after you spark-submit the uber-jar for the app.
What's the "magic" in local.ksh? Why don't you sbt assembly and do spark-submit with the uber-jar? Pozdrawiam, Jacek Laskowski ---- https://medium.com/@jaceklaskowski/ Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark Follow me at https://twitter.com/jaceklaskowski On Sun, Aug 14, 2016 at 11:52 AM, Mich Talebzadeh <mich.talebza...@gmail.com> wrote: > Thanks Jacek, > > I thought there was some dependency issue. This did the trick > > libraryDependencies += "org.apache.spark" %% "spark-core" % "2.0.0" > libraryDependencies += "org.apache.spark" %% "spark-sql" % "2.0.0" > libraryDependencies += "org.apache.spark" %% "spark-hive" % "2.0.0" % > "provided" > > I use a shell script that builds the jar file depending on type (sbt, mvn, > assembly) and submits it via spark-submit .. > > ./local.ksh -A ETL_scratchpad_dummy -T sbt > > As I understand "provided" means that the dependencies will be provided at > run-time (spark-submit) through the jar files but they are not needed at > compile time. > > Having said that am I correct that error message below > > [error] bad symbolic reference. A signature in HiveContext.class refers to > type Logging > [error] in package org.apache.spark which is not available. > [error] It may be completely missing from the current classpath, or the > version on > [error] the classpath might be incompatible with the version used when > compiling HiveContext.class. > [error] one error found > [error] (compile:compileIncremental) Compilation failed > > meant that some form of libraries incompatibility was happening at compile > time? > > Cheers > > > Dr Mich Talebzadeh > > > > LinkedIn > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw > > > > http://talebzadehmich.wordpress.com > > > Disclaimer: Use it at your own risk. Any and all responsibility for any > loss, damage or destruction of data or any other property which may arise > from relying on this email's technical content is explicitly disclaimed. The > author will in no case be liable for any monetary damages arising from such > loss, damage or destruction. > > > > > On 14 August 2016 at 19:11, Jacek Laskowski <ja...@japila.pl> wrote: >> >> Go to spark-shell and do :imports. You'll see all the imports and you >> could copy and paste them in your app. (but there are not many >> honestly and that won't help you much) >> >> HiveContext lives in spark-hive. You don't need spark-sql and >> spark-hive since the latter uses the former as a dependency (unless >> you're using types that come from the other dependencies). You don't >> need spark-core either. Make the dependencies simpler by: >> >> libraryDependencies += "org.apache.spark" %% "spark-hive" % "2.0.0" >> >> and mark it % Provided. >> >> The reason for provided is that you don't need that for uber-jar that >> you're going to spark-submit. >> >> Don't forget to reload your session of sbt you're compiling in. Unsure >> how you do it so quit your sbt session and do `sbt compile`. >> >> Ask away if you need more details. >> >> Pozdrawiam, >> Jacek Laskowski >> ---- >> https://medium.com/@jaceklaskowski/ >> Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark >> Follow me at https://twitter.com/jaceklaskowski >> >> >> On Sun, Aug 14, 2016 at 9:26 AM, Mich Talebzadeh >> <mich.talebza...@gmail.com> wrote: >> > The issue is on Spark shell this works OK >> > >> > Spark context Web UI available at http://50.140.197.217:55555 >> > Spark context available as 'sc' (master = local, app id = >> > local-1471191662017). >> > Spark session available as 'spark'. >> > Welcome to >> > ____ __ >> > / __/__ ___ _____/ /__ >> > _\ \/ _ \/ _ `/ __/ '_/ >> > /___/ .__/\_,_/_/ /_/\_\ version 2.0.0 >> > /_/ >> > Using Scala version 2.11.8 (Java HotSpot(TM) 64-Bit Server VM, Java >> > 1.8.0_77) >> > Type in expressions to have them evaluated. >> > Type :help for more information. >> > scala> import org.apache.spark.SparkContext >> > scala> import org.apache.spark.SparkConf >> > scala> import org.apache.spark.sql.Row >> > scala> import org.apache.spark.sql.hive.HiveContext >> > scala> import org.apache.spark.sql.types._ >> > scala> import org.apache.spark.sql.SparkSession >> > scala> import org.apache.spark.sql.functions._ >> > >> > The code itself >> > >> > >> > scala> val conf = new SparkConf(). >> > | setAppName("ETL_scratchpad_dummy"). >> > | set("spark.driver.allowMultipleContexts", "true"). >> > | set("enableHiveSupport","true") >> > conf: org.apache.spark.SparkConf = org.apache.spark.SparkConf@33215ffb >> > >> > scala> val sc = new SparkContext(conf) >> > sc: org.apache.spark.SparkContext = >> > org.apache.spark.SparkContext@3cbfdf5c >> > >> > scala> val HiveContext = new org.apache.spark.sql.hive.HiveContext(sc) >> > warning: there was one deprecation warning; re-run with -deprecation for >> > details >> > HiveContext: org.apache.spark.sql.hive.HiveContext = >> > org.apache.spark.sql.hive.HiveContext@2152fde5 >> > >> > scala> HiveContext.sql("use oraclehadoop") >> > res0: org.apache.spark.sql.DataFrame = [] >> > >> > I think I am getting something missing here a dependency >> > >> > >> > Dr Mich Talebzadeh >> > >> > >> > >> > LinkedIn >> > >> > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >> > >> > >> > >> > http://talebzadehmich.wordpress.com >> > >> > >> > Disclaimer: Use it at your own risk. Any and all responsibility for any >> > loss, damage or destruction of data or any other property which may >> > arise >> > from relying on this email's technical content is explicitly disclaimed. >> > The >> > author will in no case be liable for any monetary damages arising from >> > such >> > loss, damage or destruction. >> > >> > >> > >> > >> > On 14 August 2016 at 17:16, Koert Kuipers <ko...@tresata.com> wrote: >> >> >> >> HiveContext is gone >> >> >> >> SparkSession now combines functionality of SqlContext and HiveContext >> >> (if >> >> hive support is available) >> >> >> >> On Sun, Aug 14, 2016 at 12:12 PM, Mich Talebzadeh >> >> <mich.talebza...@gmail.com> wrote: >> >>> >> >>> Thanks Koert, >> >>> >> >>> I did that before as well. Anyway this is dependencies >> >>> >> >>> libraryDependencies += "org.apache.spark" %% "spark-core" % "2.0.0" >> >>> libraryDependencies += "org.apache.spark" %% "spark-sql" % "2.0.0" >> >>> libraryDependencies += "org.apache.spark" %% "spark-hive" % "2.0.0" >> >>> >> >>> >> >>> and the error >> >>> >> >>> >> >>> [info] Compiling 1 Scala source to >> >>> /data6/hduser/scala/ETL_scratchpad_dummy/target/scala-2.10/classes... >> >>> [error] >> >>> >> >>> /data6/hduser/scala/ETL_scratchpad_dummy/src/main/scala/ETL_scratchpad_dummy.scala:4: >> >>> object hive is not a member of package org.apache.spark.sql >> >>> [error] import org.apache.spark.sql.hive.HiveContext >> >>> [error] ^ >> >>> [error] >> >>> >> >>> /data6/hduser/scala/ETL_scratchpad_dummy/src/main/scala/ETL_scratchpad_dummy.scala:20: >> >>> object hive is not a member of package org.apache.spark.sql >> >>> [error] val HiveContext = new >> >>> org.apache.spark.sql.hive.HiveContext(sc) >> >>> >> >>> >> >>> >> >>> >> >>> >> >>> Dr Mich Talebzadeh >> >>> >> >>> >> >>> >> >>> LinkedIn >> >>> >> >>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >> >>> >> >>> >> >>> >> >>> http://talebzadehmich.wordpress.com >> >>> >> >>> >> >>> Disclaimer: Use it at your own risk. Any and all responsibility for >> >>> any >> >>> loss, damage or destruction of data or any other property which may >> >>> arise >> >>> from relying on this email's technical content is explicitly >> >>> disclaimed. The >> >>> author will in no case be liable for any monetary damages arising from >> >>> such >> >>> loss, damage or destruction. >> >>> >> >>> >> >>> >> >>> >> >>> On 14 August 2016 at 17:00, Koert Kuipers <ko...@tresata.com> wrote: >> >>>> >> >>>> you cannot mix spark 1 and spark 2 jars >> >>>> >> >>>> change this >> >>>> libraryDependencies += "org.apache.spark" %% "spark-hive" % "1.5.1" >> >>>> to >> >>>> libraryDependencies += "org.apache.spark" %% "spark-hive" % "2.0.0" >> >>>> >> >>>> On Sun, Aug 14, 2016 at 11:58 AM, Mich Talebzadeh >> >>>> <mich.talebza...@gmail.com> wrote: >> >>>>> >> >>>>> Hi, >> >>>>> >> >>>>> In Spark 2 I am using sbt or mvn to compile my scala program. This >> >>>>> used >> >>>>> to compile and run perfectly with Spark 1.6.1 but now it is throwing >> >>>>> error >> >>>>> >> >>>>> >> >>>>> I believe the problem is here. I have >> >>>>> >> >>>>> name := "scala" >> >>>>> version := "1.0" >> >>>>> scalaVersion := "2.11.7" >> >>>>> libraryDependencies += "org.apache.spark" %% "spark-core" % "2.0.0" >> >>>>> libraryDependencies += "org.apache.spark" %% "spark-sql" % "2.0.0" >> >>>>> libraryDependencies += "org.apache.spark" %% "spark-hive" % "1.5.1" >> >>>>> >> >>>>> However the error I am getting is >> >>>>> >> >>>>> [error] bad symbolic reference. A signature in HiveContext.class >> >>>>> refers >> >>>>> to type Logging >> >>>>> [error] in package org.apache.spark which is not available. >> >>>>> [error] It may be completely missing from the current classpath, or >> >>>>> the >> >>>>> version on >> >>>>> [error] the classpath might be incompatible with the version used >> >>>>> when >> >>>>> compiling HiveContext.class. >> >>>>> [error] one error found >> >>>>> [error] (compile:compileIncremental) Compilation failed >> >>>>> >> >>>>> >> >>>>> And this is the code >> >>>>> >> >>>>> import org.apache.spark.SparkContext >> >>>>> import org.apache.spark.SparkConf >> >>>>> import org.apache.spark.sql.Row >> >>>>> import org.apache.spark.sql.hive.HiveContext >> >>>>> import org.apache.spark.sql.types._ >> >>>>> import org.apache.spark.sql.SparkSession >> >>>>> import org.apache.spark.sql.functions._ >> >>>>> object ETL_scratchpad_dummy { >> >>>>> def main(args: Array[String]) { >> >>>>> val conf = new SparkConf(). >> >>>>> setAppName("ETL_scratchpad_dummy"). >> >>>>> set("spark.driver.allowMultipleContexts", "true"). >> >>>>> set("enableHiveSupport","true") >> >>>>> val sc = new SparkContext(conf) >> >>>>> //import sqlContext.implicits._ >> >>>>> val HiveContext = new org.apache.spark.sql.hive.HiveContext(sc) >> >>>>> HiveContext.sql("use oraclehadoop") >> >>>>> >> >>>>> >> >>>>> Anyone has come across this >> >>>>> >> >>>>> >> >>>>> >> >>>>> Dr Mich Talebzadeh >> >>>>> >> >>>>> >> >>>>> >> >>>>> LinkedIn >> >>>>> >> >>>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >> >>>>> >> >>>>> >> >>>>> >> >>>>> http://talebzadehmich.wordpress.com >> >>>>> >> >>>>> >> >>>>> Disclaimer: Use it at your own risk. Any and all responsibility for >> >>>>> any >> >>>>> loss, damage or destruction of data or any other property which may >> >>>>> arise >> >>>>> from relying on this email's technical content is explicitly >> >>>>> disclaimed. The >> >>>>> author will in no case be liable for any monetary damages arising >> >>>>> from such >> >>>>> loss, damage or destruction. >> >>>>> >> >>>>> >> >>>> >> >>>> >> >>> >> >> >> > > > --------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org