Here's the JIRA on this known issue: https://issues.apache.org/jira/browse/SPARK-1835
tl;dr: manually delete mesos-0.18.1.jar from lib_managed/jars after running sbt/sbt gen-idea. You should be able to run units inside Intellij after doing so. Doris On Tue, Jun 17, 2014 at 6:10 PM, Henry Saputra <henry.sapu...@gmail.com> wrote: > I got stuck on this one too after did git pull from master. > > Have not been able to resolve it yet =( > > > - Henry > > On Wed, Jun 11, 2014 at 6:51 AM, Yijie Shen <henry.yijies...@gmail.com> > wrote: > > Thx Qiuzhuang, the problems disappeared after I add assembly jar at the > head of list dependencies in *.iml, but while running test in Spark > SQL(SQLQuerySuite in sql-core), another two error occurs: > > > > Error 1: > > Error:scalac: > > while compiling: > /Users/yijie/code/apache.spark.master/sql/core/src/main/scala/org/apache/spark/sql/test/TestSQLContext.scala > > during phase: jvm > > library version: version 2.10.4 > > compiler version: version 2.10.4 > > reconstructed args: -Xmax-classfile-name 120 -deprecation > -P:genjavadoc:out=/Users/yijie/code/apache.spark.master/sql/core/target/java > -feature -classpath > /Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/ant-javafx.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/dt.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/javafx-doclet.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/javafx-mx.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/jconsole.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/sa-jdi.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/tools.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/charsets.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Conte… > > … > > ... > > > /Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/charsets.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/jfr.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/classes:/Users/yijie/code/apache.spark.master/lib_managed/jars/scala-library-2.10.4.jar > -Xplugin:/Users/yijie/code/apache.spark.master/lib_managed/jars/genjavadoc-plugin_2.10.4-0.5.jar > -Xplugin:/Users/yijie/code/apache.spark.master/lib_managed/jars/genjavadoc-plugin_2.10.4-0.5.jar > > last tree to typer: Literal(Constant(parquet.io.api.Converter)) > > symbol: null > > symbol definition: null > > tpe: Class(classOf[parquet.io.api.Converter]) > > symbol owners: > > context owners: object TestSQLContext -> package test > > == Enclosing template or block == > > Template( // val <local TestSQLContext>: <notype> in object > TestSQLContext, tree.tpe=org.apache.spark.sql.test.TestSQLContext.type > > "org.apache.spark.sql.SQLContext" // parents > > ValDef( > > private > > "_" > > <tpt> > > <empty> > > ) > > // 2 statements > > DefDef( // private def readResolve(): Object in object TestSQLContext > > <method> private <synthetic> > > "readResolve" > > [] > > List(Nil) > > <tpt> // tree.tpe=Object > > test.this."TestSQLContext" // object TestSQLContext in package test, > tree.tpe=org.apache.spark.sql.test.TestSQLContext.type > > ) > > DefDef( // def <init>(): org.apache.spark.sql.test.TestSQLContext.type > in object TestSQLContext > > <method> > > "<init>" > > [] > > List(Nil) > > <tpt> // tree.tpe=org.apache.spark.sql.test.TestSQLContext.type > > Block( // tree.tpe=Unit > > Apply( // def <init>(sparkContext: org.apache.spark.SparkContext): > org.apache.spark.sql.SQLContext in class SQLContext, > tree.tpe=org.apache.spark.sql.SQLContext > > TestSQLContext.super."<init>" // def <init>(sparkContext: > org.apache.spark.SparkContext): org.apache.spark.sql.SQLContext in class > SQLContext, tree.tpe=(sparkContext: > org.apache.spark.SparkContext)org.apache.spark.sql.SQLContext > > Apply( // def <init>(master: String,appName: String,conf: > org.apache.spark.SparkConf): org.apache.spark.SparkContext in class > SparkContext, tree.tpe=org.apache.spark.SparkContext > > new org.apache.spark.SparkContext."<init>" // def > <init>(master: String,appName: String,conf: org.apache.spark.SparkConf): > org.apache.spark.SparkContext in class SparkContext, tree.tpe=(master: > String, appName: String, conf: > org.apache.spark.SparkConf)org.apache.spark.SparkContext > > // 3 arguments > > "local" > > "TestSQLContext" > > Apply( // def <init>(): org.apache.spark.SparkConf in class > SparkConf, tree.tpe=org.apache.spark.SparkConf > > new org.apache.spark.SparkConf."<init>" // def <init>(): > org.apache.spark.SparkConf in class SparkConf, > tree.tpe=()org.apache.spark.SparkConf > > Nil > > ) > > ) > > ) > > () > > ) > > ) > > ) > > == Expanded type of tree == > > ConstantType(value = Constant(parquet.io.api.Converter)) > > uncaught exception during compilation: java.lang.AssertionError > > > > Error 2: > > > > Error:scalac: Error: assertion failed: List(object package$DebugNode, > object package$DebugNode) > > java.lang.AssertionError: assertion failed: List(object > package$DebugNode, object package$DebugNode) > > at > scala.reflect.internal.Symbols$Symbol.suchThat(Symbols.scala:1678) > > at > scala.reflect.internal.Symbols$ClassSymbol.companionModule0(Symbols.scala:2988) > > at > scala.reflect.internal.Symbols$ClassSymbol.companionModule(Symbols.scala:2991) > > at > scala.tools.nsc.backend.jvm.GenASM$JPlainBuilder.genClass(GenASM.scala:1371) > > at > scala.tools.nsc.backend.jvm.GenASM$AsmPhase.run(GenASM.scala:120) > > at > scala.tools.nsc.Global$Run.compileUnitsInternal(Global.scala:1583) > > at scala.tools.nsc.Global$Run.compileUnits(Global.scala:1557) > > at scala.tools.nsc.Global$Run.compileSources(Global.scala:1553) > > at scala.tools.nsc.Global$Run.compile(Global.scala:1662) > > at xsbt.CachedCompiler0.run(CompilerInterface.scala:126) > > at xsbt.CachedCompiler0.run(CompilerInterface.scala:102) > > at xsbt.CompilerInterface.run(CompilerInterface.scala:27) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:606) > > at > sbt.compiler.AnalyzingCompiler.call(AnalyzingCompiler.scala:102) > > at > sbt.compiler.AnalyzingCompiler.compile(AnalyzingCompiler.scala:48) > > at > sbt.compiler.AnalyzingCompiler.compile(AnalyzingCompiler.scala:41) > > at > org.jetbrains.jps.incremental.scala.local.IdeaIncrementalCompiler.compile(IdeaIncrementalCompiler.scala:28) > > at > org.jetbrains.jps.incremental.scala.local.LocalServer.compile(LocalServer.scala:25) > > at > org.jetbrains.jps.incremental.scala.remote.Main$.make(Main.scala:64) > > at > org.jetbrains.jps.incremental.scala.remote.Main$.nailMain(Main.scala:22) > > at > org.jetbrains.jps.incremental.scala.remote.Main.nailMain(Main.scala) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:606) > > at com.martiansoftware.nailgun.NGSession.run(NGSession.java:319) > > > > > > On Jun 11, 2014, at 11:17 AM, Qiuzhuang Lian <qiuzhuang.l...@gmail.com> > wrote: > > > >> I also run into this problem when running examples in IDEA. The issue > looks that it uses depends on too many jars and that the classpath seems to > have length limit. So I import the assembly jar and put the head of the > list dependent path and it works. > >> > >> Thanks, > >> Qiuzhuang > >> > >> > >> On Wed, Jun 11, 2014 at 10:39 AM, 申毅杰 <henry.yijies...@gmail.com> > wrote: > >> Hi All, > >> > >> I want to run ScalaTest Suite in IDEA directly, but it seems didn’t > pass the make phase before test running. > >> The problems are as follows: > >> > >> > /Users/yijie/code/apache.spark.master/core/src/main/scala/org/apache/spark/executor/MesosExecutorBackend.scala > >> Error:(44, 35) type mismatch; > >> found : org.apache.mesos.protobuf.ByteString > >> required: com.google.protobuf.ByteString > >> .setData(ByteString.copyFrom(data)) > >> ^ > >> > /Users/yijie/code/apache.spark.master/core/src/main/scala/org/apache/spark/scheduler/cluster/mesos/MesosSchedulerBackend.scala > >> Error:(119, 35) type mismatch; > >> found : org.apache.mesos.protobuf.ByteString > >> required: com.google.protobuf.ByteString > >> .setData(ByteString.copyFrom(createExecArg())) > >> ^ > >> Error:(257, 35) type mismatch; > >> found : org.apache.mesos.protobuf.ByteString > >> required: com.google.protobuf.ByteString > >> .setData(ByteString.copyFrom(task.serializedTask)) > >> ^ > >> > >> Before I run test in IDEA, I build spark through ’sbt/sbt assembly’, > >> import projects into IDEA after ’sbt/sbt gen-idea’, > >> and able to run test in Terminal ’sbt/sbt test’ > >> > >> Are there anything I leave out in order to run/debug testsuite inside > IDEA? > >> > >> Best regards, > >> Yijie > >> > > >