Yes I have seen this same error - and for team members as well - repeatedly since June. A Patrick and Cheng mentioned, the next step is to do an sbt clean
2014-11-02 19:37 GMT-08:00 Cheng Lian <lian.cs....@gmail.com>: > I often see this when I first build the whole Spark project with SBT, then > modify some code and tries to build and debug within IDEA, or vice versa. A > clean rebuild can always solve this. > > On Mon, Nov 3, 2014 at 11:28 AM, Patrick Wendell <pwend...@gmail.com> > wrote: > > > Does this happen if you clean and recompile? I've seen failures on and > > off, but haven't been able to find one that I could reproduce from a > > clean build such that we could hand it to the scala team. > > > > - Patrick > > > > On Sun, Nov 2, 2014 at 7:25 PM, Imran Rashid <im...@therashids.com> > wrote: > > > I'm finding the scala compiler crashes when I compile the spark-sql > > project > > > in sbt. This happens in both the 1.1 branch and master (full error > > > below). The other projects build fine in sbt, and everything builds > fine > > > in maven. is there some sbt option I'm forgetting? Any one else > > > experiencing this? > > > > > > Also, are there up-to-date instructions on how to do common dev tasks > in > > > both sbt & maven? I have only found these instructions on building > with > > > maven: > > > > > > http://spark.apache.org/docs/latest/building-with-maven.html > > > > > > and some general info here: > > > > > > > https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark > > > > > > but I think this doesn't walk through a lot of the steps of a typical > dev > > > cycle, eg, continuous compilation, running one test, running one main > > > class, etc. (especially since it seems like people still favor sbt for > > > dev.) If it doesn't already exist somewhere, I could try to put > > together a > > > brief doc for how to do the basics. (I'm returning to spark dev after > a > > > little hiatus myself, and I'm hitting some stumbling blocks that are > > > probably common knowledge to everyone still dealing with it all the > > time.) > > > > > > thanks, > > > Imran > > > > > > ------------------------------ > > > full crash info from sbt: > > > > > >> project sql > > > [info] Set current project to spark-sql (in build > > > file:/Users/imran/spark/spark/) > > >> compile > > > [info] Compiling 62 Scala sources to > > > /Users/imran/spark/spark/sql/catalyst/target/scala-2.10/classes... > > > [info] Compiling 45 Scala sources and 39 Java sources to > > > /Users/imran/spark/spark/sql/core/target/scala-2.10/classes... > > > [error] > > > [error] while compiling: > > > > > > /Users/imran/spark/spark/sql/core/src/main/scala/org/apache/spark/sql/types/util/DataTypeConversions.scala > > > [error] during phase: jvm > > > [error] library version: version 2.10.4 > > > [error] compiler version: version 2.10.4 > > > [error] reconstructed args: -classpath > > > > > > /Users/imran/spark/spark/sql/core/target/scala-2.10/classes:/Users/imran/spark/spark/core/target/scala-2.10/classes:/Users/imran/spark/spark/sql/catalyst/target/scala-2.10/classes:/Users/imran/spark/spark/lib_managed/jars/hadoop-client-1.0.4.jar:/Users/imran/spark/spark/lib_managed/jars/hadoop-core-1.0.4.jar:/Users/imran/spark/spark/lib_managed/jars/xmlenc-0.52.jar:/Users/imran/spark/spark/lib_managed/jars/commons-math-2.1.jar:/Users/imran/spark/spark/lib_managed/jars/commons-configuration-1.6.jar:/Users/imran/spark/spark/lib_managed/jars/commons-collections-3.2.1.jar:/Users/imran/spark/spark/lib_managed/jars/commons-lang-2.4.jar:/Users/imran/spark/spark/lib_managed/jars/commons-logging-1.1.1.jar:/Users/imran/spark/spark/lib_managed/jars/commons-digester-1.8.jar:/Users/imran/spark/spark/lib_managed/jars/commons-beanutils-1.7.0.jar:/Users/imran/spark/spark/lib_managed/jars/commons-beanutils-core-1.8.0.jar:/Users/imran/spark/spark/lib_managed/jars/commons-net-2.2.jar:/Users/imran/spark/spark/lib_managed/jars/commons-el-1.0.jar:/Users/imran/spark/spark/lib_managed/jars/hsqldb-1.8.0.10.jar:/Users/imran/spark/spark/lib_managed/jars/oro-2.0.8.jar:/Users/imran/spark/spark/lib_managed/jars/jets3t-0.7.1.jar:/Users/imran/spark/spark/lib_managed/jars/commons-httpclient-3.1.jar:/Users/imran/spark/spark/lib_managed/bundles/curator-recipes-2.4.0.jar:/Users/imran/spark/spark/lib_managed/bundles/curator-framework-2.4.0.jar:/Users/imran/spark/spark/lib_managed/bundles/curator-client-2.4.0.jar:/Users/imran/spark/spark/lib_managed/jars/zookeeper-3.4.5.jar:/Users/imran/spark/spark/lib_managed/jars/slf4j-log4j12-1.7.5.jar:/Users/imran/spark/spark/lib_managed/bundles/log4j-1.2.17.jar:/Users/imran/spark/spark/lib_managed/jars/jline-0.9.94.jar:/Users/imran/spark/spark/lib_managed/bundles/guava-14.0.1.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-plus-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/orbits/javax.transaction-1.1.1.v201105210645.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-webapp-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-xml-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-util-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-servlet-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-security-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-server-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/orbits/javax.servlet-3.0.0.v201112011016.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-continuation-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-http-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-io-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-jndi-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/orbits/javax.mail.glassfish-1.4.1.v201005082020.jar:/Users/imran/spark/spark/lib_managed/orbits/javax.activation-1.1.0.v201105071233.jar:/Users/imran/spark/spark/lib_managed/jars/commons-lang3-3.3.2.jar:/Users/imran/spark/spark/lib_managed/jars/jsr305-1.3.9.jar:/Users/imran/spark/spark/lib_managed/jars/slf4j-api-1.7.5.jar:/Users/imran/spark/spark/lib_managed/jars/jul-to-slf4j-1.7.5.jar:/Users/imran/spark/spark/lib_managed/jars/jcl-over-slf4j-1.7.5.jar:/Users/imran/spark/spark/lib_managed/bundles/compress-lzf-1.0.0.jar:/Users/imran/spark/spark/lib_managed/bundles/snappy-java-1.0.5.3.jar:/Users/imran/spark/spark/lib_managed/jars/lz4-1.2.0.jar:/Users/imran/spark/spark/lib_managed/jars/chill_2.10-0.3.6.jar:/Users/imran/spark/spark/lib_managed/jars/chill-java-0.3.6.jar:/Users/imran/spark/spark/lib_managed/bundles/kryo-2.21.jar:/Users/imran/spark/spark/lib_managed/jars/reflectasm-1.07-shaded.jar:/Users/imran/spark/spark/lib_managed/jars/minlog-1.2.jar:/Users/imran/spark/spark/lib_managed/jars/objenesis-1.2.jar:/Users/imran/spark/spark/lib_managed/bundles/akka-remote_2.10-2.2.3-shaded-protobuf.jar:/Users/imran/spark/spark/lib_managed/jars/akka-actor_2.10-2.2.3-shaded-protobuf.jar:/Users/imran/spark/spark/lib_managed/bundles/config-1.0.2.jar:/Users/imran/spark/spark/lib_managed/bundles/netty-3.6.6.Final.jar:/Users/imran/spark/spark/lib_managed/jars/protobuf-java-2.4.1-shaded.jar:/Users/imran/spark/spark/lib_managed/jars/uncommons-maths-1.2.2a.jar:/Users/imran/spark/spark/lib_managed/bundles/akka-slf4j_2.10-2.2.3-shaded-protobuf.jar:/Users/imran/spark/spark/lib_managed/jars/json4s-jackson_2.10-3.2.10.jar:/Users/imran/spark/spark/lib_managed/jars/json4s-core_2.10-3.2.10.jar:/Users/imran/spark/spark/lib_managed/jars/json4s-ast_2.10-3.2.10.jar:/Users/imran/spark/spark/lib_managed/jars/paranamer-2.6.jar:/Users/imran/spark/spark/lib_managed/jars/scalap-2.10.0.jar:/Users/imran/spark/spark/lib_managed/bundles/jackson-databind-2.3.1.jar:/Users/imran/spark/spark/lib_managed/bundles/jackson-annotations-2.3.0.jar:/Users/imran/spark/spark/lib_managed/bundles/jackson-core-2.3.1.jar:/Users/imran/spark/spark/lib_managed/jars/colt-1.2.0.jar:/Users/imran/spark/spark/lib_managed/jars/concurrent-1.3.4.jar:/Users/imran/spark/spark/lib_managed/jars/mesos-0.18.1-shaded-protobuf.jar:/Users/imran/spark/spark/lib_managed/jars/netty-all-4.0.23.Final.jar:/Users/imran/spark/spark/lib_managed/jars/stream-2.7.0.jar:/Users/imran/spark/spark/lib_managed/bundles/metrics-core-3.0.0.jar:/Users/imran/spark/spark/lib_managed/bundles/metrics-jvm-3.0.0.jar:/Users/imran/spark/spark/lib_managed/bundles/metrics-json-3.0.0.jar:/Users/imran/spark/spark/lib_managed/bundles/metrics-graphite-3.0.0.jar:/Users/imran/spark/spark/lib_managed/jars/tachyon-client-0.5.0.jar:/Users/imran/spark/spark/lib_managed/jars/tachyon-0.5.0.jar:/Users/imran/spark/spark/lib_managed/jars/commons-io-2.4.jar:/Users/imran/spark/spark/lib_managed/jars/pyrolite-2.0.1.jar:/Users/imran/spark/spark/lib_managed/jars/py4j-0.8.2.1.jar:/Users/imran/.sbt/boot/scala-2.10.4/lib/scala-compiler.jar:/Users/imran/.sbt/boot/scala-2.10.4/lib/scala-reflect.jar:/Users/imran/spark/spark/lib_managed/jars/quasiquotes_2.10-2.0.1.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-column-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-common-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-encoding-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-generator-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/commons-codec-1.5.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-hadoop-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-format-2.0.0.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-jackson-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/jackson-mapper-asl-1.9.11 > > > > > > .jar:/Users/imran/spark/spark/lib_managed/jars/jackson-core-asl-1.9.11.jar > > > -deprecation -feature > > > -P:genjavadoc:out=/Users/imran/spark/spark/sql/core/target/java > > > > > > -Xplugin:/Users/imran/spark/spark/lib_managed/jars/genjavadoc-plugin_2.10.4-0.7.jar > > > -bootclasspath > > > > > > /Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/resources.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/rt.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/sunrsasign.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/jsse.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/jce.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/charsets.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/jfr.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/classes:/Users/imran/.sbt/boot/scala-2.10.4/lib/scala-library.jar > > > -unchecked -language:postfixOps > > > [error] > > > [error] last tree to typer: > > > Literal(Constant(org.apache.spark.sql.catalyst.types.PrimitiveType)) > > > [error] symbol: null > > > [error] symbol definition: null > > > [error] tpe: > > > Class(classOf[org.apache.spark.sql.catalyst.types.PrimitiveType]) > > > [error] symbol owners: > > > [error] context owners: anonymous class anonfun$asScalaDataType$1 > > -> > > > package util > > > [error] > > > [error] == Enclosing template or block == > > > [error] > > > [error] Template( // val <local $anonfun>: <notype>, > > > tree.tpe=org.apache.spark.sql.types.util.anonfun$asScalaDataType$1 > > > [error] "scala.runtime.AbstractFunction1", "scala.Serializable" // > > parents > > > [error] ValDef( > > > [error] private > > > [error] "_" > > > [error] <tpt> > > > [error] <empty> > > > [error] ) > > > [error] // 3 statements > > > [error] DefDef( // final def apply(javaStructField: > > > org.apache.spark.sql.api.java.StructField): > > > org.apache.spark.sql.catalyst.types.StructField > > > [error] <method> final <triedcooking> > > > [error] "apply" > > > [error] [] > > > [error] // 1 parameter list > > > [error] ValDef( // javaStructField: > > > org.apache.spark.sql.api.java.StructField > > > [error] <param> <synthetic> <triedcooking> > > > [error] "javaStructField" > > > [error] <tpt> // > tree.tpe=org.apache.spark.sql.api.java.StructField > > > [error] <empty> > > > [error] ) > > > [error] <tpt> // > > > tree.tpe=org.apache.spark.sql.catalyst.types.StructField > > > [error] Apply( // def asScalaStructField(javaStructField: > > > org.apache.spark.sql.api.java.StructField): > > > org.apache.spark.sql.catalyst.types.StructField in object > > > DataTypeConversions, > > > tree.tpe=org.apache.spark.sql.catalyst.types.StructField > > > [error] DataTypeConversions.this."asScalaStructField" // def > > > asScalaStructField(javaStructField: > > > org.apache.spark.sql.api.java.StructField): > > > org.apache.spark.sql.catalyst.types.StructField in object > > > DataTypeConversions, tree.tpe=(javaStructField: > > > > > > org.apache.spark.sql.api.java.StructField)org.apache.spark.sql.catalyst.types.StructField > > > [error] "javaStructField" // javaStructField: > > > org.apache.spark.sql.api.java.StructField, > > > tree.tpe=org.apache.spark.sql.api.java.StructField > > > [error] ) > > > [error] ) > > > [error] DefDef( // final def apply(v1: Object): Object > > > [error] <method> final <bridge> > > > [error] "apply" > > > [error] [] > > > [error] // 1 parameter list > > > [error] ValDef( // v1: Object > > > [error] <param> <triedcooking> > > > [error] "v1" > > > [error] <tpt> // tree.tpe=Object > > > [error] <empty> > > > [error] ) > > > [error] <tpt> // tree.tpe=Object > > > [error] Apply( // final def apply(javaStructField: > > > org.apache.spark.sql.api.java.StructField): > > > org.apache.spark.sql.catalyst.types.StructField, > > > tree.tpe=org.apache.spark.sql.catalyst.types.StructField > > > [error] > DataTypeConversions$$anonfun$asScalaDataType$1.this."apply" > > > // final def apply(javaStructField: > > > org.apache.spark.sql.api.java.StructField): > > > org.apache.spark.sql.catalyst.types.StructField, > > tree.tpe=(javaStructField: > > > > > > org.apache.spark.sql.api.java.StructField)org.apache.spark.sql.catalyst.types.StructField > > > [error] Apply( // final def $asInstanceOf[T0 >: ? <: ?](): T0 in > > > class Object, tree.tpe=org.apache.spark.sql.api.java.StructField > > > [error] TypeApply( // final def $asInstanceOf[T0 >: ? <: ?](): > T0 > > > in class Object, tree.tpe=()org.apache.spark.sql.api.java.StructField > > > [error] "v1"."$asInstanceOf" // final def $asInstanceOf[T0 > >: ? > > > <: ?](): T0 in class Object, tree.tpe=[T0 >: ? <: ?]()T0 > > > [error] <tpt> // > > > tree.tpe=org.apache.spark.sql.api.java.StructField > > > [error] ) > > > [error] Nil > > > [error] ) > > > [error] ) > > > [error] ) > > > [error] DefDef( // def <init>(): > > > org.apache.spark.sql.types.util.anonfun$asScalaDataType$1 > > > [error] <method> <triedcooking> > > > [error] "<init>" > > > [error] [] > > > [error] List(Nil) > > > [error] <tpt> // > > > tree.tpe=org.apache.spark.sql.types.util.anonfun$asScalaDataType$1 > > > [error] Block( // tree.tpe=Unit > > > [error] Apply( // def <init>(): scala.runtime.AbstractFunction1 > in > > > class AbstractFunction1, tree.tpe=scala.runtime.AbstractFunction1 > > > [error] > > > DataTypeConversions$$anonfun$asScalaDataType$1.super."<init>" // def > > > <init>(): scala.runtime.AbstractFunction1 in class AbstractFunction1, > > > tree.tpe=()scala.runtime.AbstractFunction1 > > > [error] Nil > > > [error] ) > > > [error] () > > > [error] ) > > > [error] ) > > > [error] ) > > > [error] > > > [error] == Expanded type of tree == > > > [error] > > > [error] ConstantType( > > > [error] value = > > > Constant(org.apache.spark.sql.catalyst.types.PrimitiveType) > > > [error] ) > > > [error] > > > [error] uncaught exception during compilation: java.lang.AssertionError > > > [trace] Stack trace suppressed: run last sql/compile:compile for the > full > > > output. > > > [error] (sql/compile:compile) java.lang.AssertionError: assertion > failed: > > > List(object package$DebugNode, object package$DebugNode) > > > [error] Total time: 23 s, completed Nov 2, 2014 1:00:37 PM > > > > --------------------------------------------------------------------- > > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org > > For additional commands, e-mail: dev-h...@spark.apache.org > > > > >