Re: Tests failed after assembling the latest code from github
Thank you for your reply. After building the assembly jar, the repl test still failed. The error output is same as I post before. -- Ye Xianjin Sent with Sparrow (http://www.sparrowmailapp.com/?sig) On Tuesday, April 15, 2014 at 1:39 AM, Michael Armbrust wrote: I believe you may need an assembly jar to run the ReplSuite. sbt/sbt assembly/assembly. Michael On Mon, Apr 14, 2014 at 3:14 AM, Ye Xianjin advance...@gmail.com (mailto:advance...@gmail.com) wrote: Hi, everyone: I am new to Spark development. I download spark's latest code from github. After running sbt/sbt assembly, I began running sbt/sbt test in the spark source code dir. But it failed running the repl module test. Here are some output details. command: sbt/sbt test-only org.apache.spark.repl.* output: [info] Loading project definition from /Volumes/MacintoshHD/github/spark/project/project [info] Loading project definition from /Volumes/MacintoshHD/github/spark/project [info] Set current project to root (in build file:/Volumes/MacintoshHD/github/spark/) [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for graphx/test:testOnly [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for bagel/test:testOnly [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for streaming/test:testOnly [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for mllib/test:testOnly [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for catalyst/test:testOnly [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for core/test:testOnly [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for assembly/test:testOnly [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for sql/test:testOnly [info] ExecutorClassLoaderSuite: 2014-04-14 16:59:31.247 java[8393:1003] Unable to load realm info from SCDynamicStore [info] - child first *** FAILED *** (440 milliseconds) [info] java.lang.ClassNotFoundException: ReplFakeClass2 [info] at java.lang.ClassLoader.findClass(ClassLoader.java:364) [info] at org.apache.spark.util.ParentClassLoader.findClass(ParentClassLoader.scala:26) [info] at java.lang.ClassLoader.loadClass(ClassLoader.java:306) [info] at java.lang.ClassLoader.loadClass(ClassLoader.java:247) [info] at org.apache.spark.util.ParentClassLoader.loadClass(ParentClassLoader.scala:30) [info] at org.apache.spark.repl.ExecutorClassLoader$$anonfun$findClass$1.apply(ExecutorClassLoader.scala:57) [info] at org.apache.spark.repl.ExecutorClassLoader$$anonfun$findClass$1.apply(ExecutorClassLoader.scala:57) [info] at scala.Option.getOrElse(Option.scala:120) [info] at org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:57) [info] at java.lang.ClassLoader.loadClass(ClassLoader.java:306) [info] at java.lang.ClassLoader.loadClass(ClassLoader.java:247) [info] at org.apache.spark.repl.ExecutorClassLoaderSuite$$anonfun$1.apply$mcV$sp(ExecutorClassLoaderSuite.scala:47) [info] at org.apache.spark.repl.ExecutorClassLoaderSuite$$anonfun$1.apply(ExecutorClassLoaderSuite.scala:44) [info] at org.apache.spark.repl.ExecutorClassLoaderSuite$$anonfun$1.apply(ExecutorClassLoaderSuite.scala:44) [info] at org.scalatest.FunSuite$$anon$1.apply(FunSuite.scala:1265) [info] at org.scalatest.Suite$class.withFixture(Suite.scala:1974) [info] at org.apache.spark.repl.ExecutorClassLoaderSuite.withFixture(ExecutorClassLoaderSuite.scala:30) [info] at org.scalatest.FunSuite$class.invokeWithFixture$1(FunSuite.scala:1262) [info] at org.scalatest.FunSuite$$anonfun$runTest$1.apply(FunSuite.scala:1271) [info] at org.scalatest.FunSuite$$anonfun$runTest$1.apply(FunSuite.scala:1271) [info] at org.scalatest.SuperEngine.runTestImpl(Engine.scala:198) [info] at org.scalatest.FunSuite$class.runTest(FunSuite.scala:1271) [info] at org.apache.spark.repl.ExecutorClassLoaderSuite.runTest(ExecutorClassLoaderSuite.scala:30) [info] at org.scalatest.FunSuite$$anonfun$runTests$1.apply(FunSuite.scala:1304) [info] at org.scalatest.FunSuite$$anonfun$runTests$1.apply(FunSuite.scala:1304) [info] at org.scalatest.SuperEngine$$anonfun$org$scalatest$SuperEngine$$runTestsInBranch$1.apply(Engine.scala:260) [info] at org.scalatest.SuperEngine$$anonfun$org$scalatest$SuperEngine$$runTestsInBranch$1.apply(Engine.scala:249) [info] at scala.collection.immutable.List.foreach(List.scala:318) [info] at org.scalatest.SuperEngine.org (http://org.scalatest.SuperEngine.org) $scalatest$SuperEngine$$runTestsInBranch(Engine.scala:249) [info] at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:326) [info] at org.scalatest.FunSuite$class.runTests(FunSuite.scala:1304) [info] at
Re: Tests failed after assembling the latest code from github
Hi, I think I have found the cause of the tests failing. I have two disks on my laptop. The spark project dir is on an HDD disk while the tempdir created by google.io.Files.createTempDir is the /var/folders/5q/ ,which is on the system disk, an SSD. The ExecutorLoaderSuite test uses org.apache.spark.TestUtils.createdCompiledClass methods. The createCompiledClass method first generates the compiled class in the pwd(spark/repl), thens use renameTo to move the file. The renameTo method fails because the dest file is in a different filesystem than the source file. I modify the TestUtils.scala to first copy the file to dest then delete the original file. The tests go smoothly. Should I issue an jira about this problem? Then I can send a pr on Github. -- Ye Xianjin Sent with Sparrow (http://www.sparrowmailapp.com/?sig) On Tuesday, April 15, 2014 at 3:43 AM, Ye Xianjin wrote: well. This is very strange. I looked into ExecutorClassLoaderSuite.scala and ReplSuite.scala and made small changes to ExecutorClassLoaderSuite.scala (mostly output some internal variables). After that, when running repl test, I noticed the ReplSuite was tested first and the test result is ok. But the ExecutorClassLoaderSuite test was weird. Here is the output: [info] ExecutorClassLoaderSuite: [error] Uncaught exception when running org.apache.spark.repl.ExecutorClassLoaderSuite: java.lang.OutOfMemoryError: PermGen space [error] Uncaught exception when running org.apache.spark.repl.ExecutorClassLoaderSuite: java.lang.OutOfMemoryError: PermGen space Internal error when running tests: java.lang.OutOfMemoryError: PermGen space Exception in thread Thread-3 java.io.EOFException at java.io.ObjectInputStream$BlockDataInputStream.peekByte(ObjectInputStream.java:2577) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1297) at java.io.ObjectInputStream.readArray(ObjectInputStream.java:1685) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1323) at java.io.ObjectInputStream.readObject(ObjectInputStream.java:349) at sbt.React.react(ForkTests.scala:116) at sbt.ForkTests$$anonfun$mainTestTask$1$Acceptor$2$.run(ForkTests.scala:75) at java.lang.Thread.run(Thread.java:695) I revert my changes. The test result is same. I touched the ReplSuite.scala file (use touch command), the test order is reversed, same as the very beginning. And the output is also the same.(The result in my first post). -- Ye Xianjin Sent with Sparrow (http://www.sparrowmailapp.com/?sig) On Tuesday, April 15, 2014 at 3:14 AM, Aaron Davidson wrote: This may have something to do with running the tests on a Mac, as there is a lot of File/URI/URL stuff going on in that test which may just have happened to work if run on a Linux system (like Jenkins). Note that this suite was added relatively recently: https://github.com/apache/spark/pull/217 On Mon, Apr 14, 2014 at 12:04 PM, Ye Xianjin advance...@gmail.com (mailto:advance...@gmail.com) wrote: Thank you for your reply. After building the assembly jar, the repl test still failed. The error output is same as I post before. -- Ye Xianjin Sent with Sparrow (http://www.sparrowmailapp.com/?sig) On Tuesday, April 15, 2014 at 1:39 AM, Michael Armbrust wrote: I believe you may need an assembly jar to run the ReplSuite. sbt/sbt assembly/assembly. Michael On Mon, Apr 14, 2014 at 3:14 AM, Ye Xianjin advance...@gmail.com (mailto:advance...@gmail.com)(mailto: advance...@gmail.com (mailto:advance...@gmail.com)) wrote: Hi, everyone: I am new to Spark development. I download spark's latest code from github. After running sbt/sbt assembly, I began running sbt/sbt test in the spark source code dir. But it failed running the repl module test. Here are some output details. command: sbt/sbt test-only org.apache.spark.repl.* output: [info] Loading project definition from /Volumes/MacintoshHD/github/spark/project/project [info] Loading project definition from /Volumes/MacintoshHD/github/spark/project [info] Set current project to root (in build file:/Volumes/MacintoshHD/github/spark/) [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for graphx/test:testOnly [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for bagel/test:testOnly [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for streaming/test:testOnly [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for mllib/test:testOnly [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for catalyst/test:testOnly [info] Passed: Total 0, Failed 0, Errors 0, Passed 0 [info] No tests to run for