What do people think of running the Big Data Benchmark <https://amplab.cs.berkeley.edu/benchmark/> (repo <https://github.com/amplab/benchmark>) as part of preparing every new release of Spark?
We'd run it just for Spark and effectively use it as another type of test to track any performance progress or regressions from release to release. Would doing such a thing be valuable? Do we already have a way of benchmarking Spark performance that we use regularly? Nick