Hi Mich,

Yeah, you don't have to worry about it...and that's why you're asking
these questions ;-)

Pozdrawiam,
Jacek Laskowski
----
https://medium.com/@jaceklaskowski/
Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark
Follow me at https://twitter.com/jaceklaskowski


On Sun, Aug 14, 2016 at 12:06 PM, Mich Talebzadeh
<mich.talebza...@gmail.com> wrote:
> The magic does all that(including compiling and submitting with the jar
> file. It is flexible as it does all this for any Sala program. it creates
> sub-directories, compiles, submits etc so I don't have to worry about it.
>
> HTH
>
> Dr Mich Talebzadeh
>
>
>
> LinkedIn
> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>
>
>
> http://talebzadehmich.wordpress.com
>
>
> Disclaimer: Use it at your own risk. Any and all responsibility for any
> loss, damage or destruction of data or any other property which may arise
> from relying on this email's technical content is explicitly disclaimed. The
> author will in no case be liable for any monetary damages arising from such
> loss, damage or destruction.
>
>
>
>
> On 14 August 2016 at 20:01, Jacek Laskowski <ja...@japila.pl> wrote:
>>
>> Hi,
>>
>> You should have all the deps being "provided" since they're provided
>> by spark infra after you spark-submit the uber-jar for the app.
>>
>> What's the "magic" in local.ksh? Why don't you sbt assembly and do
>> spark-submit with the uber-jar?
>>
>> Pozdrawiam,
>> Jacek Laskowski
>> ----
>> https://medium.com/@jaceklaskowski/
>> Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark
>> Follow me at https://twitter.com/jaceklaskowski
>>
>>
>> On Sun, Aug 14, 2016 at 11:52 AM, Mich Talebzadeh
>> <mich.talebza...@gmail.com> wrote:
>> > Thanks Jacek,
>> >
>> > I thought there was some dependency issue. This did the trick
>> >
>> > libraryDependencies += "org.apache.spark" %% "spark-core" % "2.0.0"
>> > libraryDependencies += "org.apache.spark" %% "spark-sql" % "2.0.0"
>> > libraryDependencies += "org.apache.spark" %% "spark-hive" % "2.0.0" %
>> > "provided"
>> >
>> > I use a shell script that builds the jar file depending on type (sbt,
>> > mvn,
>> > assembly)  and submits it via spark-submit ..
>> >
>> > ./local.ksh -A ETL_scratchpad_dummy -T sbt
>> >
>> > As I understand "provided" means that the dependencies will be provided
>> > at
>> > run-time (spark-submit) through the jar files but they are not needed at
>> > compile time.
>> >
>> > Having said that am I correct that error message below
>> >
>> > [error] bad symbolic reference. A signature in HiveContext.class refers
>> > to
>> > type Logging
>> > [error] in package org.apache.spark which is not available.
>> > [error] It may be completely missing from the current classpath, or the
>> > version on
>> > [error] the classpath might be incompatible with the version used when
>> > compiling HiveContext.class.
>> > [error] one error found
>> > [error] (compile:compileIncremental) Compilation failed
>> >
>> > meant that some form of libraries incompatibility was happening at
>> > compile
>> > time?
>> >
>> > Cheers
>> >
>> >
>> > Dr Mich Talebzadeh
>> >
>> >
>> >
>> > LinkedIn
>> >
>> > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>> >
>> >
>> >
>> > http://talebzadehmich.wordpress.com
>> >
>> >
>> > Disclaimer: Use it at your own risk. Any and all responsibility for any
>> > loss, damage or destruction of data or any other property which may
>> > arise
>> > from relying on this email's technical content is explicitly disclaimed.
>> > The
>> > author will in no case be liable for any monetary damages arising from
>> > such
>> > loss, damage or destruction.
>> >
>> >
>> >
>> >
>> > On 14 August 2016 at 19:11, Jacek Laskowski <ja...@japila.pl> wrote:
>> >>
>> >> Go to spark-shell and do :imports. You'll see all the imports and you
>> >> could copy and paste them in your app. (but there are not many
>> >> honestly and that won't help you much)
>> >>
>> >> HiveContext lives in spark-hive. You don't need spark-sql and
>> >> spark-hive since the latter uses the former as a dependency (unless
>> >> you're using types that come from the other dependencies). You don't
>> >> need spark-core either. Make the dependencies simpler by:
>> >>
>> >> libraryDependencies += "org.apache.spark" %% "spark-hive" % "2.0.0"
>> >>
>> >> and mark it % Provided.
>> >>
>> >> The reason for provided is that you don't need that for uber-jar that
>> >> you're going to spark-submit.
>> >>
>> >> Don't forget to reload your session of sbt you're compiling in. Unsure
>> >> how you do it so quit your sbt session and do `sbt compile`.
>> >>
>> >> Ask away if you need more details.
>> >>
>> >> Pozdrawiam,
>> >> Jacek Laskowski
>> >> ----
>> >> https://medium.com/@jaceklaskowski/
>> >> Mastering Apache Spark 2.0 http://bit.ly/mastering-apache-spark
>> >> Follow me at https://twitter.com/jaceklaskowski
>> >>
>> >>
>> >> On Sun, Aug 14, 2016 at 9:26 AM, Mich Talebzadeh
>> >> <mich.talebza...@gmail.com> wrote:
>> >> > The issue is on Spark shell this works OK
>> >> >
>> >> > Spark context Web UI available at http://50.140.197.217:55555
>> >> > Spark context available as 'sc' (master = local, app id =
>> >> > local-1471191662017).
>> >> > Spark session available as 'spark'.
>> >> > Welcome to
>> >> >       ____              __
>> >> >      / __/__  ___ _____/ /__
>> >> >     _\ \/ _ \/ _ `/ __/  '_/
>> >> >    /___/ .__/\_,_/_/ /_/\_\   version 2.0.0
>> >> >       /_/
>> >> > Using Scala version 2.11.8 (Java HotSpot(TM) 64-Bit Server VM, Java
>> >> > 1.8.0_77)
>> >> > Type in expressions to have them evaluated.
>> >> > Type :help for more information.
>> >> > scala> import org.apache.spark.SparkContext
>> >> > scala> import org.apache.spark.SparkConf
>> >> > scala> import org.apache.spark.sql.Row
>> >> > scala> import org.apache.spark.sql.hive.HiveContext
>> >> > scala> import org.apache.spark.sql.types._
>> >> > scala> import org.apache.spark.sql.SparkSession
>> >> > scala> import org.apache.spark.sql.functions._
>> >> >
>> >> > The code itself
>> >> >
>> >> >
>> >> > scala>   val conf = new SparkConf().
>> >> >      |                setAppName("ETL_scratchpad_dummy").
>> >> >      |                set("spark.driver.allowMultipleContexts",
>> >> > "true").
>> >> >      |                set("enableHiveSupport","true")
>> >> > conf: org.apache.spark.SparkConf =
>> >> > org.apache.spark.SparkConf@33215ffb
>> >> >
>> >> > scala>   val sc = new SparkContext(conf)
>> >> > sc: org.apache.spark.SparkContext =
>> >> > org.apache.spark.SparkContext@3cbfdf5c
>> >> >
>> >> > scala>   val HiveContext = new
>> >> > org.apache.spark.sql.hive.HiveContext(sc)
>> >> > warning: there was one deprecation warning; re-run with -deprecation
>> >> > for
>> >> > details
>> >> > HiveContext: org.apache.spark.sql.hive.HiveContext =
>> >> > org.apache.spark.sql.hive.HiveContext@2152fde5
>> >> >
>> >> > scala>   HiveContext.sql("use oraclehadoop")
>> >> > res0: org.apache.spark.sql.DataFrame = []
>> >> >
>> >> > I think I am getting something missing here a dependency
>> >> >
>> >> >
>> >> > Dr Mich Talebzadeh
>> >> >
>> >> >
>> >> >
>> >> > LinkedIn
>> >> >
>> >> >
>> >> > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>> >> >
>> >> >
>> >> >
>> >> > http://talebzadehmich.wordpress.com
>> >> >
>> >> >
>> >> > Disclaimer: Use it at your own risk. Any and all responsibility for
>> >> > any
>> >> > loss, damage or destruction of data or any other property which may
>> >> > arise
>> >> > from relying on this email's technical content is explicitly
>> >> > disclaimed.
>> >> > The
>> >> > author will in no case be liable for any monetary damages arising
>> >> > from
>> >> > such
>> >> > loss, damage or destruction.
>> >> >
>> >> >
>> >> >
>> >> >
>> >> > On 14 August 2016 at 17:16, Koert Kuipers <ko...@tresata.com> wrote:
>> >> >>
>> >> >> HiveContext is gone
>> >> >>
>> >> >> SparkSession now combines functionality of SqlContext and
>> >> >> HiveContext
>> >> >> (if
>> >> >> hive support is available)
>> >> >>
>> >> >> On Sun, Aug 14, 2016 at 12:12 PM, Mich Talebzadeh
>> >> >> <mich.talebza...@gmail.com> wrote:
>> >> >>>
>> >> >>> Thanks Koert,
>> >> >>>
>> >> >>> I did that before as well. Anyway this is dependencies
>> >> >>>
>> >> >>> libraryDependencies += "org.apache.spark" %% "spark-core" % "2.0.0"
>> >> >>> libraryDependencies += "org.apache.spark" %% "spark-sql" % "2.0.0"
>> >> >>> libraryDependencies += "org.apache.spark" %% "spark-hive" % "2.0.0"
>> >> >>>
>> >> >>>
>> >> >>> and the error
>> >> >>>
>> >> >>>
>> >> >>> [info] Compiling 1 Scala source to
>> >> >>>
>> >> >>> /data6/hduser/scala/ETL_scratchpad_dummy/target/scala-2.10/classes...
>> >> >>> [error]
>> >> >>>
>> >> >>>
>> >> >>> /data6/hduser/scala/ETL_scratchpad_dummy/src/main/scala/ETL_scratchpad_dummy.scala:4:
>> >> >>> object hive is not a member of package org.apache.spark.sql
>> >> >>> [error] import org.apache.spark.sql.hive.HiveContext
>> >> >>> [error]                             ^
>> >> >>> [error]
>> >> >>>
>> >> >>>
>> >> >>> /data6/hduser/scala/ETL_scratchpad_dummy/src/main/scala/ETL_scratchpad_dummy.scala:20:
>> >> >>> object hive is not a member of package org.apache.spark.sql
>> >> >>> [error]   val HiveContext = new
>> >> >>> org.apache.spark.sql.hive.HiveContext(sc)
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>> Dr Mich Talebzadeh
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>> LinkedIn
>> >> >>>
>> >> >>>
>> >> >>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>> http://talebzadehmich.wordpress.com
>> >> >>>
>> >> >>>
>> >> >>> Disclaimer: Use it at your own risk. Any and all responsibility for
>> >> >>> any
>> >> >>> loss, damage or destruction of data or any other property which may
>> >> >>> arise
>> >> >>> from relying on this email's technical content is explicitly
>> >> >>> disclaimed. The
>> >> >>> author will in no case be liable for any monetary damages arising
>> >> >>> from
>> >> >>> such
>> >> >>> loss, damage or destruction.
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>> On 14 August 2016 at 17:00, Koert Kuipers <ko...@tresata.com>
>> >> >>> wrote:
>> >> >>>>
>> >> >>>> you cannot mix spark 1 and spark 2 jars
>> >> >>>>
>> >> >>>> change this
>> >> >>>> libraryDependencies += "org.apache.spark" %% "spark-hive" %
>> >> >>>> "1.5.1"
>> >> >>>> to
>> >> >>>> libraryDependencies += "org.apache.spark" %% "spark-hive" %
>> >> >>>> "2.0.0"
>> >> >>>>
>> >> >>>> On Sun, Aug 14, 2016 at 11:58 AM, Mich Talebzadeh
>> >> >>>> <mich.talebza...@gmail.com> wrote:
>> >> >>>>>
>> >> >>>>> Hi,
>> >> >>>>>
>> >> >>>>> In Spark 2 I am using sbt or mvn to compile my scala program.
>> >> >>>>> This
>> >> >>>>> used
>> >> >>>>> to compile and run perfectly with Spark 1.6.1 but now it is
>> >> >>>>> throwing
>> >> >>>>> error
>> >> >>>>>
>> >> >>>>>
>> >> >>>>> I believe the problem is here. I have
>> >> >>>>>
>> >> >>>>> name := "scala"
>> >> >>>>> version := "1.0"
>> >> >>>>> scalaVersion := "2.11.7"
>> >> >>>>> libraryDependencies += "org.apache.spark" %% "spark-core" %
>> >> >>>>> "2.0.0"
>> >> >>>>> libraryDependencies += "org.apache.spark" %% "spark-sql" %
>> >> >>>>> "2.0.0"
>> >> >>>>> libraryDependencies += "org.apache.spark" %% "spark-hive" %
>> >> >>>>> "1.5.1"
>> >> >>>>>
>> >> >>>>> However the error I am getting is
>> >> >>>>>
>> >> >>>>> [error] bad symbolic reference. A signature in HiveContext.class
>> >> >>>>> refers
>> >> >>>>> to type Logging
>> >> >>>>> [error] in package org.apache.spark which is not available.
>> >> >>>>> [error] It may be completely missing from the current classpath,
>> >> >>>>> or
>> >> >>>>> the
>> >> >>>>> version on
>> >> >>>>> [error] the classpath might be incompatible with the version used
>> >> >>>>> when
>> >> >>>>> compiling HiveContext.class.
>> >> >>>>> [error] one error found
>> >> >>>>> [error] (compile:compileIncremental) Compilation failed
>> >> >>>>>
>> >> >>>>>
>> >> >>>>> And this is the code
>> >> >>>>>
>> >> >>>>> import org.apache.spark.SparkContext
>> >> >>>>> import org.apache.spark.SparkConf
>> >> >>>>> import org.apache.spark.sql.Row
>> >> >>>>> import org.apache.spark.sql.hive.HiveContext
>> >> >>>>> import org.apache.spark.sql.types._
>> >> >>>>> import org.apache.spark.sql.SparkSession
>> >> >>>>> import org.apache.spark.sql.functions._
>> >> >>>>> object ETL_scratchpad_dummy {
>> >> >>>>>   def main(args: Array[String]) {
>> >> >>>>>   val conf = new SparkConf().
>> >> >>>>>                setAppName("ETL_scratchpad_dummy").
>> >> >>>>>                set("spark.driver.allowMultipleContexts", "true").
>> >> >>>>>                set("enableHiveSupport","true")
>> >> >>>>>   val sc = new SparkContext(conf)
>> >> >>>>>   //import sqlContext.implicits._
>> >> >>>>>   val HiveContext = new org.apache.spark.sql.hive.HiveContext(sc)
>> >> >>>>>   HiveContext.sql("use oraclehadoop")
>> >> >>>>>
>> >> >>>>>
>> >> >>>>> Anyone has come across this
>> >> >>>>>
>> >> >>>>>
>> >> >>>>>
>> >> >>>>> Dr Mich Talebzadeh
>> >> >>>>>
>> >> >>>>>
>> >> >>>>>
>> >> >>>>> LinkedIn
>> >> >>>>>
>> >> >>>>>
>> >> >>>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>> >> >>>>>
>> >> >>>>>
>> >> >>>>>
>> >> >>>>> http://talebzadehmich.wordpress.com
>> >> >>>>>
>> >> >>>>>
>> >> >>>>> Disclaimer: Use it at your own risk. Any and all responsibility
>> >> >>>>> for
>> >> >>>>> any
>> >> >>>>> loss, damage or destruction of data or any other property which
>> >> >>>>> may
>> >> >>>>> arise
>> >> >>>>> from relying on this email's technical content is explicitly
>> >> >>>>> disclaimed. The
>> >> >>>>> author will in no case be liable for any monetary damages arising
>> >> >>>>> from such
>> >> >>>>> loss, damage or destruction.
>> >> >>>>>
>> >> >>>>>
>> >> >>>>
>> >> >>>>
>> >> >>>
>> >> >>
>> >> >
>> >
>> >
>
>

---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to