What do people think of running the Big Data Benchmark
<https://amplab.cs.berkeley.edu/benchmark/> (repo
<https://github.com/amplab/benchmark>) as part of preparing every new
release of Spark?

We'd run it just for Spark and effectively use it as another type of test
to track any performance progress or regressions from release to release.

Would doing such a thing be valuable? Do we already have a way of
benchmarking Spark performance that we use regularly?

Nick

Reply via email to