I got a clean version of the master branch, and do the steps as follows:

export MAVEN_OPTS="-Xmx2g -XX:MaxPermSize=512M -XX:ReservedCodeCacheSize=512m”
mvn -U -Dhadoop.version=2.2.0 -DskipTests clean package

after these steps, I open the project in IDEA through pom.xml in the root 
folder, but while run the same test SQLQuerySuite in sql-core, the two errors 
above still occurs, any ideas? 

On Jun 11, 2014, at 10:16 PM, Qiuzhuang Lian <qiuzhuang.l...@gmail.com> wrote:

> I run into this issue too today via 'mvn install -DskipTests' command today, 
> then I issue a mvn clean and rebuild and it works.
> 
> Thanks,
> Qiuzhuang
> 
> 
> On Wed, Jun 11, 2014 at 9:51 PM, Yijie Shen <henry.yijies...@gmail.com> wrote:
> Thx Qiuzhuang, the problems disappeared after I add assembly jar at the head 
> of list dependencies in *.iml, but while running test in Spark 
> SQL(SQLQuerySuite in sql-core), another two error occurs:
> 
> Error 1: 
> Error:scalac: 
>      while compiling: 
> /Users/yijie/code/apache.spark.master/sql/core/src/main/scala/org/apache/spark/sql/test/TestSQLContext.scala
>         during phase: jvm
>      library version: version 2.10.4
>     compiler version: version 2.10.4
>   reconstructed args: -Xmax-classfile-name 120 -deprecation 
> -P:genjavadoc:out=/Users/yijie/code/apache.spark.master/sql/core/target/java 
> -feature -classpath 
> /Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/ant-javafx.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/dt.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/javafx-doclet.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/javafx-mx.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/jconsole.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/sa-jdi.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/lib/tools.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/charsets.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Conte…
> …
> ...
> /Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/charsets.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/jfr.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/classes:/Users/yijie/code/apache.spark.master/lib_managed/jars/scala-library-2.10.4.jar
>  
> -Xplugin:/Users/yijie/code/apache.spark.master/lib_managed/jars/genjavadoc-plugin_2.10.4-0.5.jar
>  
> -Xplugin:/Users/yijie/code/apache.spark.master/lib_managed/jars/genjavadoc-plugin_2.10.4-0.5.jar
>   last tree to typer: Literal(Constant(parquet.io.api.Converter))
>               symbol: null
>    symbol definition: null
>                  tpe: Class(classOf[parquet.io.api.Converter])
>        symbol owners: 
>       context owners: object TestSQLContext -> package test
> == Enclosing template or block ==
> Template( // val <local TestSQLContext>: <notype> in object TestSQLContext, 
> tree.tpe=org.apache.spark.sql.test.TestSQLContext.type
>   "org.apache.spark.sql.SQLContext" // parents
>   ValDef(
>     private
>     "_"
>     <tpt>
>     <empty>
>   )
>   // 2 statements
>   DefDef( // private def readResolve(): Object in object TestSQLContext
>     <method> private <synthetic>
>     "readResolve"
>     []
>     List(Nil)
>     <tpt> // tree.tpe=Object
>     test.this."TestSQLContext" // object TestSQLContext in package test, 
> tree.tpe=org.apache.spark.sql.test.TestSQLContext.type
>   )
>   DefDef( // def <init>(): org.apache.spark.sql.test.TestSQLContext.type in 
> object TestSQLContext
>     <method>
>     "<init>"
>     []
>     List(Nil)
>     <tpt> // tree.tpe=org.apache.spark.sql.test.TestSQLContext.type
>     Block( // tree.tpe=Unit
>       Apply( // def <init>(sparkContext: org.apache.spark.SparkContext): 
> org.apache.spark.sql.SQLContext in class SQLContext, 
> tree.tpe=org.apache.spark.sql.SQLContext
>         TestSQLContext.super."<init>" // def <init>(sparkContext: 
> org.apache.spark.SparkContext): org.apache.spark.sql.SQLContext in class 
> SQLContext, tree.tpe=(sparkContext: 
> org.apache.spark.SparkContext)org.apache.spark.sql.SQLContext
>         Apply( // def <init>(master: String,appName: String,conf: 
> org.apache.spark.SparkConf): org.apache.spark.SparkContext in class 
> SparkContext, tree.tpe=org.apache.spark.SparkContext
>           new org.apache.spark.SparkContext."<init>" // def <init>(master: 
> String,appName: String,conf: org.apache.spark.SparkConf): 
> org.apache.spark.SparkContext in class SparkContext, tree.tpe=(master: 
> String, appName: String, conf: 
> org.apache.spark.SparkConf)org.apache.spark.SparkContext
>           // 3 arguments
>           "local"
>           "TestSQLContext"
>           Apply( // def <init>(): org.apache.spark.SparkConf in class 
> SparkConf, tree.tpe=org.apache.spark.SparkConf
>             new org.apache.spark.SparkConf."<init>" // def <init>(): 
> org.apache.spark.SparkConf in class SparkConf, 
> tree.tpe=()org.apache.spark.SparkConf
>             Nil
>           )
>         )
>       )
>       ()
>     )
>   )
> )
> == Expanded type of tree ==
> ConstantType(value = Constant(parquet.io.api.Converter))
> uncaught exception during compilation: java.lang.AssertionError
> 
> Error 2:
> 
> Error:scalac: Error: assertion failed: List(object package$DebugNode, object 
> package$DebugNode)
> java.lang.AssertionError: assertion failed: List(object package$DebugNode, 
> object package$DebugNode)
>       at scala.reflect.internal.Symbols$Symbol.suchThat(Symbols.scala:1678)
>       at 
> scala.reflect.internal.Symbols$ClassSymbol.companionModule0(Symbols.scala:2988)
>       at 
> scala.reflect.internal.Symbols$ClassSymbol.companionModule(Symbols.scala:2991)
>       at 
> scala.tools.nsc.backend.jvm.GenASM$JPlainBuilder.genClass(GenASM.scala:1371)
>       at scala.tools.nsc.backend.jvm.GenASM$AsmPhase.run(GenASM.scala:120)
>       at scala.tools.nsc.Global$Run.compileUnitsInternal(Global.scala:1583)
>       at scala.tools.nsc.Global$Run.compileUnits(Global.scala:1557)
>       at scala.tools.nsc.Global$Run.compileSources(Global.scala:1553)
>       at scala.tools.nsc.Global$Run.compile(Global.scala:1662)
>       at xsbt.CachedCompiler0.run(CompilerInterface.scala:126)
>       at xsbt.CachedCompiler0.run(CompilerInterface.scala:102)
>       at xsbt.CompilerInterface.run(CompilerInterface.scala:27)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:606)
>       at sbt.compiler.AnalyzingCompiler.call(AnalyzingCompiler.scala:102)
>       at sbt.compiler.AnalyzingCompiler.compile(AnalyzingCompiler.scala:48)
>       at sbt.compiler.AnalyzingCompiler.compile(AnalyzingCompiler.scala:41)
>       at 
> org.jetbrains.jps.incremental.scala.local.IdeaIncrementalCompiler.compile(IdeaIncrementalCompiler.scala:28)
>       at 
> org.jetbrains.jps.incremental.scala.local.LocalServer.compile(LocalServer.scala:25)
>       at org.jetbrains.jps.incremental.scala.remote.Main$.make(Main.scala:64)
>       at 
> org.jetbrains.jps.incremental.scala.remote.Main$.nailMain(Main.scala:22)
>       at org.jetbrains.jps.incremental.scala.remote.Main.nailMain(Main.scala)
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:606)
>       at com.martiansoftware.nailgun.NGSession.run(NGSession.java:319)
> 
> 
> On Jun 11, 2014, at 11:17 AM, Qiuzhuang Lian <qiuzhuang.l...@gmail.com> wrote:
> 
>> I also run into this problem when running examples in IDEA. The issue looks 
>> that it uses depends on too many jars and that the classpath seems to have 
>> length limit. So I import the assembly jar and put the head of the list 
>> dependent path and it works.
>> 
>> Thanks,
>> Qiuzhuang
>> 
>> 
>> On Wed, Jun 11, 2014 at 10:39 AM, 申毅杰 <henry.yijies...@gmail.com> wrote:
>> Hi All,
>> 
>> I want to run ScalaTest Suite in IDEA directly, but it seems didn’t pass the 
>> make phase before test running.
>> The problems are as follows:
>> 
>> /Users/yijie/code/apache.spark.master/core/src/main/scala/org/apache/spark/executor/MesosExecutorBackend.scala
>> Error:(44, 35) type mismatch;
>>  found   : org.apache.mesos.protobuf.ByteString
>>  required: com.google.protobuf.ByteString
>>       .setData(ByteString.copyFrom(data))
>>                                   ^
>> /Users/yijie/code/apache.spark.master/core/src/main/scala/org/apache/spark/scheduler/cluster/mesos/MesosSchedulerBackend.scala
>> Error:(119, 35) type mismatch;
>>  found   : org.apache.mesos.protobuf.ByteString
>>  required: com.google.protobuf.ByteString
>>       .setData(ByteString.copyFrom(createExecArg()))
>>                                   ^
>> Error:(257, 35) type mismatch;
>>  found   : org.apache.mesos.protobuf.ByteString
>>  required: com.google.protobuf.ByteString
>>       .setData(ByteString.copyFrom(task.serializedTask))
>>                                   ^
>> 
>> Before I run test in IDEA, I build spark through ’sbt/sbt assembly’,
>> import projects into IDEA after ’sbt/sbt gen-idea’,
>> and able to run test in Terminal ’sbt/sbt test’
>> 
>> Are there anything I leave out in order to run/debug testsuite inside IDEA?
>> 
>> Best regards,
>> Yijie
>> 
> 
> 

Reply via email to