[ 
https://issues.apache.org/jira/browse/SPARK-6171?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Or updated SPARK-6171:
-----------------------------
    Description: 
I ran `build/sbt clean assembly` and then started the Spark shell, then I hit 
this huge stack trace. I didn't enable hive in my build, but I wasn't planning 
on using SQL either. I can still run Spark jobs no problem, but we probably 
shouldn't be throwing this on a clean build.

{code}
15/03/04 14:09:15 INFO SparkILoop: Created spark context..
Spark context available as sc.
java.lang.NoClassDefFoundError: org/apache/hadoop/hive/conf/HiveConf
        at java.lang.Class.getDeclaredConstructors0(Native Method)
        at java.lang.Class.privateGetDeclaredConstructors(Class.java:2493)
        at java.lang.Class.getConstructor0(Class.java:2803)
        at java.lang.Class.getConstructor(Class.java:1718)
        at 
org.apache.spark.repl.SparkILoop.createSQLContext(SparkILoop.scala:1026)
        at $iwC$$iwC.<init>(<console>:9)
        at $iwC.<init>(<console>:18)
        at <init>(<console>:20)
        at .<init>(<console>:24)
        at .<clinit>(<console>)
        at .<init>(<console>:7)
        at .<clinit>(<console>)
        at $print(<console>)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at 
org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065)
        at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkI
{code}

  was:
I ran `build/sbt clean assembly` and then started the Spark shell, clean and 
simple, then I hit this huge stack trace. I can still run Spark jobs no 
problem, but we probably shouldn't be throwing this on a clean build.

{code}
15/03/04 14:09:15 INFO SparkILoop: Created spark context..
Spark context available as sc.
java.lang.NoClassDefFoundError: org/apache/hadoop/hive/conf/HiveConf
        at java.lang.Class.getDeclaredConstructors0(Native Method)
        at java.lang.Class.privateGetDeclaredConstructors(Class.java:2493)
        at java.lang.Class.getConstructor0(Class.java:2803)
        at java.lang.Class.getConstructor(Class.java:1718)
        at 
org.apache.spark.repl.SparkILoop.createSQLContext(SparkILoop.scala:1026)
        at $iwC$$iwC.<init>(<console>:9)
        at $iwC.<init>(<console>:18)
        at <init>(<console>:20)
        at .<init>(<console>:24)
        at .<clinit>(<console>)
        at .<init>(<console>:7)
        at .<clinit>(<console>)
        at $print(<console>)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at 
org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065)
        at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkI
{code}


> No class def found for HiveConf in Spark shell
> ----------------------------------------------
>
>                 Key: SPARK-6171
>                 URL: https://issues.apache.org/jira/browse/SPARK-6171
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Shell, SQL
>    Affects Versions: 1.3.0
>            Reporter: Andrew Or
>            Assignee: Michael Armbrust
>            Priority: Blocker
>
> I ran `build/sbt clean assembly` and then started the Spark shell, then I hit 
> this huge stack trace. I didn't enable hive in my build, but I wasn't 
> planning on using SQL either. I can still run Spark jobs no problem, but we 
> probably shouldn't be throwing this on a clean build.
> {code}
> 15/03/04 14:09:15 INFO SparkILoop: Created spark context..
> Spark context available as sc.
> java.lang.NoClassDefFoundError: org/apache/hadoop/hive/conf/HiveConf
>       at java.lang.Class.getDeclaredConstructors0(Native Method)
>       at java.lang.Class.privateGetDeclaredConstructors(Class.java:2493)
>       at java.lang.Class.getConstructor0(Class.java:2803)
>       at java.lang.Class.getConstructor(Class.java:1718)
>       at 
> org.apache.spark.repl.SparkILoop.createSQLContext(SparkILoop.scala:1026)
>       at $iwC$$iwC.<init>(<console>:9)
>       at $iwC.<init>(<console>:18)
>       at <init>(<console>:20)
>       at .<init>(<console>:24)
>       at .<clinit>(<console>)
>       at .<init>(<console>:7)
>       at .<clinit>(<console>)
>       at $print(<console>)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:606)
>       at 
> org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065)
>       at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkI
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to