spark-hive is excluded when using --packages, because it can be included in
the spark-assembly by adding -Phive during mvn package or sbt assembly.

Best,
Burak

On Tue, Jul 7, 2015 at 8:06 AM, Hao Ren <inv...@gmail.com> wrote:

> I want to add spark-hive as a dependence to submit my job, but it seems
> that
> spark-submit can not resolve it.
>
> $ ./bin/spark-submit \
> → --packages
>
> org.apache.spark:spark-hive_2.10:1.4.0,org.postgresql:postgresql:9.3-1103-jdbc3,joda-time:joda-time:2.8.1
> \
> → --class fr.leboncoin.etl.jobs.dwh.AdStateTraceDWHTransform \
> → --master spark://localhost:7077 \
>
> Ivy Default Cache set to: /home/invkrh/.ivy2/cache
> The jars for the packages stored in: /home/invkrh/.ivy2/jars
> https://repository.jboss.org/nexus/content/repositories/releases/ added
> as a
> remote repository with the name: repo-1
> :: loading settings :: url =
>
> jar:file:/home/invkrh/workspace/scala/spark/assembly/target/scala-2.10/spark-assembly-1.4.0-SNAPSHOT-hadoop2.2.0.jar!/org/apache/ivy/core/settings/ivysettings.xml
> org.apache.spark#spark-hive_2.10 added as a dependency
> org.postgresql#postgresql added as a dependency
> joda-time#joda-time added as a dependency
> :: resolving dependencies :: org.apache.spark#spark-submit-parent;1.0
>         confs: [default]
>         found org.postgresql#postgresql;9.3-1103-jdbc3 in local-m2-cache
>         found joda-time#joda-time;2.8.1 in central
> :: resolution report :: resolve 139ms :: artifacts dl 3ms
>         :: modules in use:
>         joda-time#joda-time;2.8.1 from central in [default]
>         org.postgresql#postgresql;9.3-1103-jdbc3 from local-m2-cache in
> [default]
>
> ---------------------------------------------------------------------
>         |                  |            modules            ||   artifacts
>  |
>         |       conf       | number| search|dwnlded|evicted||
> number|dwnlded|
>
> ---------------------------------------------------------------------
>         |      default     |   2   |   0   |   0   |   0   ||   2   |   0
>  |
>
> ---------------------------------------------------------------------
> :: retrieving :: org.apache.spark#spark-submit-parent
>         confs: [default]
>         0 artifacts copied, 2 already retrieved (0kB/6ms)
> Exception in thread "main" java.lang.NoClassDefFoundError:
> org/apache/spark/sql/hive/HiveContext
>         at java.lang.Class.forName0(Native Method)
>         at java.lang.Class.forName(Class.java:348)
>         at
>
> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:633)
>         at
> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:169)
>         at
> org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:192)
>         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:111)
>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> Caused by: java.lang.ClassNotFoundException:
> org.apache.spark.sql.hive.HiveContext
>         at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
>         at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>         at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>         ... 7 more
> Using Spark's default log4j profile:
> org/apache/spark/log4j-defaults.properties
> 15/07/07 16:57:59 INFO Utils: Shutdown hook called
>
> Any help is appreciated. Thank you.
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/spark-submit-can-not-resolve-spark-hive-2-10-tp23695.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to