This blog outlines a few things that make Spark faster than MapReduce - https://databricks.com/blog/2014/10/10/spark-petabyte-sort.html
On Fri, Aug 7, 2015 at 9:13 AM, Muler <mulugeta.abe...@gmail.com> wrote: > Consider the classic word count application over a 4 node cluster with a > sizable working data. What makes Spark ran faster than MapReduce > considering that Spark also has to write to disk during shuffle? >