I think that @holdenk's *spark-testing-base* project publishes some of
these test classes as well as some helper classes for testing streaming
jobs: https://github.com/holdenk/spark-testing-base
On Thu, May 21, 2015 at 10:39 PM, Reynold Xin r...@databricks.com wrote:
It is just 15 lines of code
see discussions about Spark not really liking multiple contexts in the
same JVM
Speaking of this - is there a standard way of writing unit tests that
require a SparkContext?
We've ended up copying out the code of SharedSparkContext to our own
testing hierarchy, but it occurs to me someone
It is just 15 lines of code to copy, isn't it?
On Thu, May 21, 2015 at 7:46 PM, Nathan Kronenfeld
nkronenfeld@uncharted.software wrote:
see discussions about Spark not really liking multiple contexts in the
same JVM
Speaking of this - is there a standard way of writing unit tests that