One thing is you need to do a "maven package" before you run tests.
The "local-cluster" tests depend on Spark already being packaged.

- Patrick

On Wed, Oct 29, 2014 at 10:02 AM, Niklas Wilcke
<1wil...@informatik.uni-hamburg.de> wrote:
> Hi Sean,
>
> thanks for your reply. The tests still don't work. I focused on the
> mllib and core tests and made some observations.
>
> The core tests seems to fail because of my german locale. Some tests are
> locale dependend like the
> UtilsSuite.scala
>  - "string formatting of time durations" - checks for locale dependend
> seperators like "." and ","
>  - "isBindCollision" - checks for the locale dependend exception message
>
> In the MLlib it seems to be just one source of failure. The same
> Exception I described in my first mail appears several times in
> different tests.
> The reason for all the similar failures is the line 29 in
> LocalClusterSparkContext.scala.
> When I change the line
> .setMaster("local-cluster[2, 1, 512]")
> to
> .setMaster("local")
> all tests run without a failure. The local-cluster mode seems to be the
> reason for the failure. I tried some different configurations like
> [1,1,512], [2,1,1024] etc. but couldn't get the tests run without a failure.
>
> Could this be a configuration issue?
>
> On 28.10.2014 19:03, Sean Owen wrote:
>> On Tue, Oct 28, 2014 at 6:18 PM, Niklas Wilcke
>> <1wil...@informatik.uni-hamburg.de> wrote:
>>> 1. via dev/run-tests script
>>>     This script executes all tests and take several hours to finish.
>>> Some tests failed but I can't say which of them. Should this really take
>>> that long? Can I specify to run only MLlib tests?
>> Yes, running all tests takes a long long time. It does print which
>> tests failed, and you can see the errors in the test output.
>>
>> Did you read 
>> http://spark.apache.org/docs/latest/building-with-maven.html#spark-tests-in-maven
>> ? This shows how to run just one test suite.
>>
>> In any Maven project you can try things like "mvn test -pl [module]"
>> to run just one module's tests.
> Yes I tried that as described below at point 2.
>>> 2. directly via maven
>>> I did the following described in the docs [0].
>>>
>>> export MAVEN_OPTS="-Xmx2g -XX:MaxPermSize=512M
>>> -XX:ReservedCodeCacheSize=512m"
>>> mvn -Pyarn -Phadoop-2.3 -DskipTests -Phive clean package
>>> mvn -Pyarn -Phadoop-2.3 -Phive test
>>>
>>> This also doesn't work.
>>> Why do I have to package spark bevore running the tests?
>> What doesn't work?
>> Some tests use the built assembly, which requires packaging.
> I get the same Exceptions as in every other way.
>>> 3. via sbt
>>> I tried the following. I freshly cloned spark and checked out the tag
>>> v1.1.0-rc4.
>>>
>>> sbt/sbt "project mllib" test
>>>
>>> and get the following exception in several cluster tests.
>>>
>>> [info] - task size should be small in both training and prediction ***
>>> FAILED ***
>> This just looks like a flaky test failure; I'd try again.
>>
> I don't think so. I tried for several times now in several different ways.
>
> Thanks,
> Niklas
>
>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> For additional commands, e-mail: dev-h...@spark.apache.org
>

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org

Reply via email to