Great! Congratulations! 

-- 
Nan Zhu


On Friday, October 10, 2014 at 11:19 AM, Mridul Muralidharan wrote:

> Brilliant stuff ! Congrats all :-)
> This is indeed really heartening news !
> 
> Regards,
> Mridul
> 
> 
> On Fri, Oct 10, 2014 at 8:24 PM, Matei Zaharia <matei.zaha...@gmail.com 
> (mailto:matei.zaha...@gmail.com)> wrote:
> > Hi folks,
> > 
> > I interrupt your regularly scheduled user / dev list to bring you some 
> > pretty cool news for the project, which is that we've been able to use 
> > Spark to break MapReduce's 100 TB and 1 PB sort records, sorting data 3x 
> > faster on 10x fewer nodes. There's a detailed writeup at 
> > http://databricks.com/blog/2014/10/10/spark-breaks-previous-large-scale-sort-record.html.
> >  Summary: while Hadoop MapReduce held last year's 100 TB world record by 
> > sorting 100 TB in 72 minutes on 2100 nodes, we sorted it in 23 minutes on 
> > 206 nodes; and we also scaled up to sort 1 PB in 234 minutes.
> > 
> > I want to thank Reynold Xin for leading this effort over the past few 
> > weeks, along with Parviz Deyhim, Xiangrui Meng, Aaron Davidson and Ali 
> > Ghodsi. In addition, we'd really like to thank Amazon's EC2 team for 
> > providing the machines to make this possible. Finally, this result would of 
> > course not be possible without the many many other contributions, testing 
> > and feature requests from throughout the community.
> > 
> > For an engine to scale from these multi-hour petabyte batch jobs down to 
> > 100-millisecond streaming and interactive queries is quite uncommon, and 
> > it's thanks to all of you folks that we are able to make this happen.
> > 
> > Matei
> > ---------------------------------------------------------------------
> > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org 
> > (mailto:dev-unsubscr...@spark.apache.org)
> > For additional commands, e-mail: dev-h...@spark.apache.org 
> > (mailto:dev-h...@spark.apache.org)
> > 
> 
> 
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org 
> (mailto:user-unsubscr...@spark.apache.org)
> For additional commands, e-mail: user-h...@spark.apache.org 
> (mailto:user-h...@spark.apache.org)
> 
> 


Reply via email to