Hi,

I am running my graphx application on Spark 1.2.0(11 nodes cluster), has 
requested 30GB memory per node and 100 cores for around 1GB input dataset(5 
million vertices graph).

But the error below always happen…

Is there anyone could give me some points? 

(BTW, the overall edge/vertex RDDs will reach more than 100GB during graph 
computation, and another version of my application can work well on the same 
dataset while it need much less memory during computation)

Thanks in advance!!!


15/01/29 18:05:08 ERROR ContextCleaner: Error cleaning broadcast 60
java.util.concurrent.TimeoutException: Futures timed out after [30 seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
        at 
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
        at scala.concurrent.Await$.result(package.scala:107)
        at 
org.apache.spark.storage.BlockManagerMaster.removeBroadcast(BlockManagerMaster.scala:137)
        at 
org.apache.spark.broadcast.TorrentBroadcast$.unpersist(TorrentBroadcast.scala:227)
        at 
org.apache.spark.broadcast.TorrentBroadcastFactory.unbroadcast(TorrentBroadcastFactory.scala:45)
        at 
org.apache.spark.broadcast.BroadcastManager.unbroadcast(BroadcastManager.scala:66)
        at 
org.apache.spark.ContextCleaner.doCleanupBroadcast(ContextCleaner.scala:185)
        at 
org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:147)
        at 
org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:138)
        at scala.Option.foreach(Option.scala:236)
        at 
org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply$mcV$sp(ContextCleaner.scala:138)
        at 
org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
        at 
org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
        at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1460)
        at 
org.apache.spark.ContextCleaner.org$apache$spark$ContextCleaner$$keepCleaning(ContextCleaner.scala:133)
        at org.apache.spark.ContextCleaner$$anon$3.run(ContextCleaner.scala:65)
[Stage 91:===================>                                      (2 + 4) / 
6]15/01/29 18:08:15 ERROR SparkDeploySchedulerBackend: Asked to remove 
non-existent executor 0
[Stage 93:================================>                      (29 + 20) / 
49]15/01/29 23:47:03 ERROR TaskSchedulerImpl: Lost executor 9 on 
small11-tap1.common.lip6.fr: remote Akka client disassociated
[Stage 83:>   (1 + 0) / 6][Stage 86:>   (0 + 1) / 2][Stage 88:>   (0 + 2) / 
8]15/01/29 23:47:06 ERROR SparkDeploySchedulerBackend: Asked to remove 
non-existent executor 9
[Stage 83:===============>  (5 + 1) / 6][Stage 88:=============>   (9 + 2) / 
11]15/01/29 23:57:30 ERROR TaskSchedulerImpl: Lost executor 8 on 
small10-tap1.common.lip6.fr: remote Akka client disassociated
15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove 
non-existent executor 8
15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove 
non-existent executor 8

Best,
Yifan LI





Reply via email to