Thanks, Sonal. But it seems to be an error happened when “cleaning broadcast”?
BTW, what is the timeout of “[30 seconds]”? can I increase it? Best, Yifan LI > On 02 Feb 2015, at 11:12, Sonal Goyal <sonalgoy...@gmail.com> wrote: > > That may be the cause of your issue. Take a look at the tuning guide[1] and > maybe also profile your application. See if you can reuse your objects. > > 1. http://spark.apache.org/docs/latest/tuning.html > <http://spark.apache.org/docs/latest/tuning.html> > > > Best Regards, > Sonal > Founder, Nube Technologies <http://www.nubetech.co/> > > <http://in.linkedin.com/in/sonalgoyal> > > > > On Sat, Jan 31, 2015 at 4:21 AM, Yifan LI <iamyifa...@gmail.com > <mailto:iamyifa...@gmail.com>> wrote: > Yes, I think so, esp. for a pregel application… have any suggestion? > > Best, > Yifan LI > > > > > >> On 30 Jan 2015, at 22:25, Sonal Goyal <sonalgoy...@gmail.com >> <mailto:sonalgoy...@gmail.com>> wrote: >> >> Is your code hitting frequent garbage collection? >> >> Best Regards, >> Sonal >> Founder, Nube Technologies <http://www.nubetech.co/> >> >> <http://in.linkedin.com/in/sonalgoyal> >> >> >> >> On Fri, Jan 30, 2015 at 7:52 PM, Yifan LI <iamyifa...@gmail.com >> <mailto:iamyifa...@gmail.com>> wrote: >> >>> >>> >>> Hi, >>> >>> I am running my graphx application on Spark 1.2.0(11 nodes cluster), has >>> requested 30GB memory per node and 100 cores for around 1GB input dataset(5 >>> million vertices graph). >>> >>> But the error below always happen… >>> >>> Is there anyone could give me some points? >>> >>> (BTW, the overall edge/vertex RDDs will reach more than 100GB during graph >>> computation, and another version of my application can work well on the >>> same dataset while it need much less memory during computation) >>> >>> Thanks in advance!!! >>> >>> >>> 15/01/29 18:05:08 ERROR ContextCleaner: Error cleaning broadcast 60 >>> java.util.concurrent.TimeoutException: Futures timed out after [30 seconds] >>> at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219) >>> at >>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) >>> at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107) >>> at >>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53) >>> at scala.concurrent.Await$.result(package.scala:107) >>> at >>> org.apache.spark.storage.BlockManagerMaster.removeBroadcast(BlockManagerMaster.scala:137) >>> at >>> org.apache.spark.broadcast.TorrentBroadcast$.unpersist(TorrentBroadcast.scala:227) >>> at >>> org.apache.spark.broadcast.TorrentBroadcastFactory.unbroadcast(TorrentBroadcastFactory.scala:45) >>> at >>> org.apache.spark.broadcast.BroadcastManager.unbroadcast(BroadcastManager.scala:66) >>> at >>> org.apache.spark.ContextCleaner.doCleanupBroadcast(ContextCleaner.scala:185) >>> at >>> org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:147) >>> at >>> org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:138) >>> at scala.Option.foreach(Option.scala:236) >>> at >>> org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply$mcV$sp(ContextCleaner.scala:138) >>> at >>> org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134) >>> at >>> org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134) >>> at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1460) >>> at org.apache.spark.ContextCleaner.org >>> <http://org.apache.spark.contextcleaner.org/>$apache$spark$ContextCleaner$$keepCleaning(ContextCleaner.scala:133) >>> at org.apache.spark.ContextCleaner$$anon$3.run(ContextCleaner.scala:65) >>> [Stage 91:===================> (2 + 4) >>> / 6]15/01/29 18:08:15 ERROR SparkDeploySchedulerBackend: Asked to remove >>> non-existent executor 0 >>> [Stage 93:================================> (29 + 20) >>> / 49]15/01/29 23:47:03 ERROR TaskSchedulerImpl: Lost executor 9 on >>> small11-tap1.common.lip6.fr <http://small11-tap1.common.lip6.fr/>: remote >>> Akka client disassociated >>> [Stage 83:> (1 + 0) / 6][Stage 86:> (0 + 1) / 2][Stage 88:> (0 + 2) / >>> 8]15/01/29 23:47:06 ERROR SparkDeploySchedulerBackend: Asked to remove >>> non-existent executor 9 >>> [Stage 83:===============> (5 + 1) / 6][Stage 88:=============> (9 + 2) >>> / 11]15/01/29 23:57:30 ERROR TaskSchedulerImpl: Lost executor 8 on >>> small10-tap1.common.lip6.fr <http://small10-tap1.common.lip6.fr/>: remote >>> Akka client disassociated >>> 15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove >>> non-existent executor 8 >>> 15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove >>> non-existent executor 8 >>> >>> Best, >>> Yifan LI >>> >>> >>> >>> >>> >> >> > >