Makes sense. Having high determinism in these tests would make Jenkins build stable.
On Mon, May 11, 2015 at 1:08 PM, Andrew Or <and...@databricks.com> wrote: > Hi Ted, > > Yes, those two options can be useful, but in general I think the standard > to set is that tests should never fail. It's actually the worst if tests > fail sometimes but not others, because we can't reproduce them > deterministically. Using -M and -A actually tolerates flaky tests to a > certain extent, and I would prefer to instead increase the determinism in > these tests. > > -Andrew > > 2015-05-08 17:56 GMT-07:00 Ted Yu <yuzhih...@gmail.com>: > > Andrew: >> Do you think the -M and -A options described here can be used in test >> runs ? >> http://scalatest.org/user_guide/using_the_runner >> >> Cheers >> >> On Wed, May 6, 2015 at 5:41 PM, Andrew Or <and...@databricks.com> wrote: >> >>> Dear all, >>> >>> I'm sure you have all noticed that the Spark tests have been fairly >>> unstable recently. I wanted to share a tool that I use to track which >>> tests >>> have been failing most often in order to prioritize fixing these flaky >>> tests. >>> >>> Here is an output of the tool. This spreadsheet reports the top 10 failed >>> tests this week (ending yesterday 5/5): >>> >>> https://docs.google.com/spreadsheets/d/1Iv_UDaTFGTMad1sOQ_s4ddWr6KD3PuFIHmTSzL7LSb4 >>> >>> It is produced by a small project: >>> https://github.com/andrewor14/spark-test-failures >>> >>> I have been filing JIRAs on flaky tests based on this tool. Hopefully we >>> can collectively stabilize the build a little more as we near the release >>> for Spark 1.4. >>> >>> -Andrew >>> >> >> >