Hi Sean, thanks for your reply. The tests still don't work. I focused on the mllib and core tests and made some observations.
The core tests seems to fail because of my german locale. Some tests are locale dependend like the UtilsSuite.scala - "string formatting of time durations" - checks for locale dependend seperators like "." and "," - "isBindCollision" - checks for the locale dependend exception message In the MLlib it seems to be just one source of failure. The same Exception I described in my first mail appears several times in different tests. The reason for all the similar failures is the line 29 in LocalClusterSparkContext.scala. When I change the line .setMaster("local-cluster[2, 1, 512]") to .setMaster("local") all tests run without a failure. The local-cluster mode seems to be the reason for the failure. I tried some different configurations like [1,1,512], [2,1,1024] etc. but couldn't get the tests run without a failure. Could this be a configuration issue? On 28.10.2014 19:03, Sean Owen wrote: > On Tue, Oct 28, 2014 at 6:18 PM, Niklas Wilcke > <1wil...@informatik.uni-hamburg.de> wrote: >> 1. via dev/run-tests script >> This script executes all tests and take several hours to finish. >> Some tests failed but I can't say which of them. Should this really take >> that long? Can I specify to run only MLlib tests? > Yes, running all tests takes a long long time. It does print which > tests failed, and you can see the errors in the test output. > > Did you read > http://spark.apache.org/docs/latest/building-with-maven.html#spark-tests-in-maven > ? This shows how to run just one test suite. > > In any Maven project you can try things like "mvn test -pl [module]" > to run just one module's tests. Yes I tried that as described below at point 2. >> 2. directly via maven >> I did the following described in the docs [0]. >> >> export MAVEN_OPTS="-Xmx2g -XX:MaxPermSize=512M >> -XX:ReservedCodeCacheSize=512m" >> mvn -Pyarn -Phadoop-2.3 -DskipTests -Phive clean package >> mvn -Pyarn -Phadoop-2.3 -Phive test >> >> This also doesn't work. >> Why do I have to package spark bevore running the tests? > What doesn't work? > Some tests use the built assembly, which requires packaging. I get the same Exceptions as in every other way. >> 3. via sbt >> I tried the following. I freshly cloned spark and checked out the tag >> v1.1.0-rc4. >> >> sbt/sbt "project mllib" test >> >> and get the following exception in several cluster tests. >> >> [info] - task size should be small in both training and prediction *** >> FAILED *** > This just looks like a flaky test failure; I'd try again. > I don't think so. I tried for several times now in several different ways. Thanks, Niklas --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org For additional commands, e-mail: dev-h...@spark.apache.org