Thanks Xiangrui, How to check and make sure the data is distributed evenly? Thanks again. On Wed, May 21, 2014 at 2:17 PM, Xiangrui Meng [via Apache Spark User List] <ml-node+s1001560n6187...@n3.nabble.com> wrote: > Many OutOfMemoryErrors in the log. Is your data distributed evenly? > -Xiangrui > > On Wed, May 21, 2014 at 11:23 AM, yxzhao <[hidden email]> wrote: > >> I run the pagerank example processing a large data set, 5GB in size, using >> 48 >> machines. The job got stuck at the time point: 14/05/20 21:32:17, as the >> attached log shows. It was stuck there for more than 10 hours and then I >> killed it at last. But I did not find any information explaining why it >> was >> stuck. Any suggestions? Thanks. >> >> >> Spark_OK_48_pagerank.log >> >> <http://apache-spark-user-list.1001560.n3.nabble.com/file/n6185/Spark_OK_48_pagerank.log> >> >> >> >> -- >> View this message in context: >> http://apache-spark-user-list.1001560.n3.nabble.com/Job-Processing-Large-Data-Set-Got-Stuck-tp6185.html >> Sent from the Apache Spark User List mailing list archive at Nabble.com. > > > ________________________________ > If you reply to this email, your message will be added to the discussion > below: > http://apache-spark-user-list.1001560.n3.nabble.com/Job-Processing-Large-Data-Set-Got-Stuck-tp6185p6187.html > To unsubscribe from Job Processing Large Data Set Got Stuck, click here. > NAML
-- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Job-Processing-Large-Data-Set-Got-Stuck-tp6185p6189.html Sent from the Apache Spark User List mailing list archive at Nabble.com.