Re: [Graphx Spark] Error of Lost executor and TimeoutException

2015-02-02 Thread Sonal Goyal
That may be the cause of your issue. Take a look at the tuning guide[1] and
maybe also profile your application. See if you can reuse your objects.

1. http://spark.apache.org/docs/latest/tuning.html


Best Regards,
Sonal
Founder, Nube Technologies http://www.nubetech.co

http://in.linkedin.com/in/sonalgoyal



On Sat, Jan 31, 2015 at 4:21 AM, Yifan LI iamyifa...@gmail.com wrote:

 Yes, I think so, esp. for a pregel application… have any suggestion?

 Best,
 Yifan LI





 On 30 Jan 2015, at 22:25, Sonal Goyal sonalgoy...@gmail.com wrote:

 Is your code hitting frequent garbage collection?

 Best Regards,
 Sonal
 Founder, Nube Technologies http://www.nubetech.co/

 http://in.linkedin.com/in/sonalgoyal



 On Fri, Jan 30, 2015 at 7:52 PM, Yifan LI iamyifa...@gmail.com wrote:




 Hi,

 I am running my graphx application on Spark 1.2.0(11 nodes cluster), has
 requested 30GB memory per node and 100 cores for around 1GB input dataset(5
 million vertices graph).

 But the error below always happen…

 Is there anyone could give me some points?

 (BTW, the overall edge/vertex RDDs will reach more than 100GB during
 graph computation, and another version of my application can work well on
 the same dataset while it need much less memory during computation)

 Thanks in advance!!!


 15/01/29 18:05:08 ERROR ContextCleaner: Error cleaning broadcast 60
 java.util.concurrent.TimeoutException: Futures timed out after [30
 seconds]
 at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
 at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
 at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
 at
 scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
 at scala.concurrent.Await$.result(package.scala:107)
 at
 org.apache.spark.storage.BlockManagerMaster.removeBroadcast(BlockManagerMaster.scala:137)
 at
 org.apache.spark.broadcast.TorrentBroadcast$.unpersist(TorrentBroadcast.scala:227)
 at
 org.apache.spark.broadcast.TorrentBroadcastFactory.unbroadcast(TorrentBroadcastFactory.scala:45)
 at
 org.apache.spark.broadcast.BroadcastManager.unbroadcast(BroadcastManager.scala:66)
 at
 org.apache.spark.ContextCleaner.doCleanupBroadcast(ContextCleaner.scala:185)
 at
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:147)
 at
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:138)
 at scala.Option.foreach(Option.scala:236)
 at
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply$mcV$sp(ContextCleaner.scala:138)
 at
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
 at
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
 at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1460)
 at org.apache.spark.ContextCleaner.org
 http://org.apache.spark.contextcleaner.org/
 $apache$spark$ContextCleaner$$keepCleaning(ContextCleaner.scala:133)
 at org.apache.spark.ContextCleaner$$anon$3.run(ContextCleaner.scala:65)
 [Stage 91:===  (2 +
 4) / 6]15/01/29 18:08:15 ERROR SparkDeploySchedulerBackend: Asked to remove
 non-existent executor 0
 [Stage 93:  (29 +
 20) / 49]15/01/29 23:47:03 ERROR TaskSchedulerImpl: Lost executor 9 on
 small11-tap1.common.lip6.fr: remote Akka client disassociated
 [Stage 83:   (1 + 0) / 6][Stage 86:   (0 + 1) / 2][Stage 88:   (0 + 2)
 / 8]15/01/29 23:47:06 ERROR SparkDeploySchedulerBackend: Asked to remove
 non-existent executor 9
 [Stage 83:===  (5 + 1) / 6][Stage 88:=   (9 +
 2) / 11]15/01/29 23:57:30 ERROR TaskSchedulerImpl: Lost executor 8 on
 small10-tap1.common.lip6.fr: remote Akka client disassociated
 15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove
 non-existent executor 8
 15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove
 non-existent executor 8

 Best,
 Yifan LI











Re: [Graphx Spark] Error of Lost executor and TimeoutException

2015-02-02 Thread Yifan LI
Thanks, Sonal.

But it seems to be an error happened when “cleaning broadcast”? 

BTW, what is the timeout of “[30 seconds]”? can I increase it?



Best,
Yifan LI





 On 02 Feb 2015, at 11:12, Sonal Goyal sonalgoy...@gmail.com wrote:
 
 That may be the cause of your issue. Take a look at the tuning guide[1] and 
 maybe also profile your application. See if you can reuse your objects. 
 
 1. http://spark.apache.org/docs/latest/tuning.html 
 http://spark.apache.org/docs/latest/tuning.html
 
 
 Best Regards,
 Sonal
 Founder, Nube Technologies http://www.nubetech.co/ 
 
  http://in.linkedin.com/in/sonalgoyal
 
 
 
 On Sat, Jan 31, 2015 at 4:21 AM, Yifan LI iamyifa...@gmail.com 
 mailto:iamyifa...@gmail.com wrote:
 Yes, I think so, esp. for a pregel application… have any suggestion?
 
 Best,
 Yifan LI
 
 
 
 
 
 On 30 Jan 2015, at 22:25, Sonal Goyal sonalgoy...@gmail.com 
 mailto:sonalgoy...@gmail.com wrote:
 
 Is your code hitting frequent garbage collection? 
 
 Best Regards,
 Sonal
 Founder, Nube Technologies http://www.nubetech.co/ 
 
  http://in.linkedin.com/in/sonalgoyal
 
 
 
 On Fri, Jan 30, 2015 at 7:52 PM, Yifan LI iamyifa...@gmail.com 
 mailto:iamyifa...@gmail.com wrote:
 
 
 
 Hi,
 
 I am running my graphx application on Spark 1.2.0(11 nodes cluster), has 
 requested 30GB memory per node and 100 cores for around 1GB input dataset(5 
 million vertices graph).
 
 But the error below always happen…
 
 Is there anyone could give me some points? 
 
 (BTW, the overall edge/vertex RDDs will reach more than 100GB during graph 
 computation, and another version of my application can work well on the 
 same dataset while it need much less memory during computation)
 
 Thanks in advance!!!
 
 
 15/01/29 18:05:08 ERROR ContextCleaner: Error cleaning broadcast 60
 java.util.concurrent.TimeoutException: Futures timed out after [30 seconds]
 at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
 at 
 scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
 at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
 at 
 scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
 at scala.concurrent.Await$.result(package.scala:107)
 at 
 org.apache.spark.storage.BlockManagerMaster.removeBroadcast(BlockManagerMaster.scala:137)
 at 
 org.apache.spark.broadcast.TorrentBroadcast$.unpersist(TorrentBroadcast.scala:227)
 at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.unbroadcast(TorrentBroadcastFactory.scala:45)
 at 
 org.apache.spark.broadcast.BroadcastManager.unbroadcast(BroadcastManager.scala:66)
 at 
 org.apache.spark.ContextCleaner.doCleanupBroadcast(ContextCleaner.scala:185)
 at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:147)
 at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:138)
 at scala.Option.foreach(Option.scala:236)
 at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply$mcV$sp(ContextCleaner.scala:138)
 at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
 at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
 at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1460)
 at org.apache.spark.ContextCleaner.org 
 http://org.apache.spark.contextcleaner.org/$apache$spark$ContextCleaner$$keepCleaning(ContextCleaner.scala:133)
 at org.apache.spark.ContextCleaner$$anon$3.run(ContextCleaner.scala:65)
 [Stage 91:===  (2 + 4) 
 / 6]15/01/29 18:08:15 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 0
 [Stage 93:  (29 + 20) 
 / 49]15/01/29 23:47:03 ERROR TaskSchedulerImpl: Lost executor 9 on 
 small11-tap1.common.lip6.fr http://small11-tap1.common.lip6.fr/: remote 
 Akka client disassociated
 [Stage 83:   (1 + 0) / 6][Stage 86:   (0 + 1) / 2][Stage 88:   (0 + 2) / 
 8]15/01/29 23:47:06 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 9
 [Stage 83:===  (5 + 1) / 6][Stage 88:=   (9 + 2) 
 / 11]15/01/29 23:57:30 ERROR TaskSchedulerImpl: Lost executor 8 on 
 small10-tap1.common.lip6.fr http://small10-tap1.common.lip6.fr/: remote 
 Akka client disassociated
 15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 8
 15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 8
 
 Best,
 Yifan LI
 
 
 
 
 
 
 
 
 



Re: [Graphx Spark] Error of Lost executor and TimeoutException

2015-02-02 Thread Yifan LI
I think this broadcast cleaning(memory block remove?) timeout exception was 
caused by:

15/02/02 11:48:49 ERROR TaskSchedulerImpl: Lost executor 13 on 
small18-tap1.common.lip6.fr: remote Akka client disassociated
15/02/02 11:48:49 ERROR SparkDeploySchedulerBackend: Asked to remove 
non-existent executor 13
15/02/02 11:48:49 ERROR SparkDeploySchedulerBackend: Asked to remove 
non-existent executor 13

Anyone has points on this?


Best,
Yifan LI





 On 02 Feb 2015, at 11:47, Yifan LI iamyifa...@gmail.com wrote:
 
 Thanks, Sonal.
 
 But it seems to be an error happened when “cleaning broadcast”? 
 
 BTW, what is the timeout of “[30 seconds]”? can I increase it?
 
 
 
 Best,
 Yifan LI
 
 
 
 
 
 On 02 Feb 2015, at 11:12, Sonal Goyal sonalgoy...@gmail.com 
 mailto:sonalgoy...@gmail.com wrote:
 
 That may be the cause of your issue. Take a look at the tuning guide[1] and 
 maybe also profile your application. See if you can reuse your objects. 
 
 1. http://spark.apache.org/docs/latest/tuning.html 
 http://spark.apache.org/docs/latest/tuning.html
 
 
 Best Regards,
 Sonal
 Founder, Nube Technologies http://www.nubetech.co/ 
 
  http://in.linkedin.com/in/sonalgoyal
 
 
 
 On Sat, Jan 31, 2015 at 4:21 AM, Yifan LI iamyifa...@gmail.com 
 mailto:iamyifa...@gmail.com wrote:
 Yes, I think so, esp. for a pregel application… have any suggestion?
 
 Best,
 Yifan LI
 
 
 
 
 
 On 30 Jan 2015, at 22:25, Sonal Goyal sonalgoy...@gmail.com 
 mailto:sonalgoy...@gmail.com wrote:
 
 Is your code hitting frequent garbage collection? 
 
 Best Regards,
 Sonal
 Founder, Nube Technologies http://www.nubetech.co/ 
 
  http://in.linkedin.com/in/sonalgoyal
 
 
 
 On Fri, Jan 30, 2015 at 7:52 PM, Yifan LI iamyifa...@gmail.com 
 mailto:iamyifa...@gmail.com wrote:
 
 
 
 Hi,
 
 I am running my graphx application on Spark 1.2.0(11 nodes cluster), has 
 requested 30GB memory per node and 100 cores for around 1GB input 
 dataset(5 million vertices graph).
 
 But the error below always happen…
 
 Is there anyone could give me some points? 
 
 (BTW, the overall edge/vertex RDDs will reach more than 100GB during graph 
 computation, and another version of my application can work well on the 
 same dataset while it need much less memory during computation)
 
 Thanks in advance!!!
 
 
 15/01/29 18:05:08 ERROR ContextCleaner: Error cleaning broadcast 60
 java.util.concurrent.TimeoutException: Futures timed out after [30 seconds]
at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
at 
 scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
at 
 scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
at scala.concurrent.Await$.result(package.scala:107)
at 
 org.apache.spark.storage.BlockManagerMaster.removeBroadcast(BlockManagerMaster.scala:137)
at 
 org.apache.spark.broadcast.TorrentBroadcast$.unpersist(TorrentBroadcast.scala:227)
at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.unbroadcast(TorrentBroadcastFactory.scala:45)
at 
 org.apache.spark.broadcast.BroadcastManager.unbroadcast(BroadcastManager.scala:66)
at 
 org.apache.spark.ContextCleaner.doCleanupBroadcast(ContextCleaner.scala:185)
at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:147)
at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:138)
at scala.Option.foreach(Option.scala:236)
at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply$mcV$sp(ContextCleaner.scala:138)
at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1460)
at org.apache.spark.ContextCleaner.org 
 http://org.apache.spark.contextcleaner.org/$apache$spark$ContextCleaner$$keepCleaning(ContextCleaner.scala:133)
at org.apache.spark.ContextCleaner$$anon$3.run(ContextCleaner.scala:65)
 [Stage 91:===  (2 + 
 4) / 6]15/01/29 18:08:15 ERROR SparkDeploySchedulerBackend: Asked to 
 remove non-existent executor 0
 [Stage 93:  (29 + 20) 
 / 49]15/01/29 23:47:03 ERROR TaskSchedulerImpl: Lost executor 9 on 
 small11-tap1.common.lip6.fr http://small11-tap1.common.lip6.fr/: remote 
 Akka client disassociated
 [Stage 83:   (1 + 0) / 6][Stage 86:   (0 + 1) / 2][Stage 88:   (0 + 2) 
 / 8]15/01/29 23:47:06 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 9
 [Stage 83:===  (5 + 1) / 6][Stage 

Re: [Graphx Spark] Error of Lost executor and TimeoutException

2015-01-30 Thread Sonal Goyal
Is your code hitting frequent garbage collection?

Best Regards,
Sonal
Founder, Nube Technologies http://www.nubetech.co

http://in.linkedin.com/in/sonalgoyal



On Fri, Jan 30, 2015 at 7:52 PM, Yifan LI iamyifa...@gmail.com wrote:




 Hi,

 I am running my graphx application on Spark 1.2.0(11 nodes cluster), has
 requested 30GB memory per node and 100 cores for around 1GB input dataset(5
 million vertices graph).

 But the error below always happen…

 Is there anyone could give me some points?

 (BTW, the overall edge/vertex RDDs will reach more than 100GB during graph
 computation, and another version of my application can work well on the
 same dataset while it need much less memory during computation)

 Thanks in advance!!!


 15/01/29 18:05:08 ERROR ContextCleaner: Error cleaning broadcast 60
 java.util.concurrent.TimeoutException: Futures timed out after [30 seconds]
 at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
 at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
 at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
 at
 scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
 at scala.concurrent.Await$.result(package.scala:107)
 at
 org.apache.spark.storage.BlockManagerMaster.removeBroadcast(BlockManagerMaster.scala:137)
 at
 org.apache.spark.broadcast.TorrentBroadcast$.unpersist(TorrentBroadcast.scala:227)
 at
 org.apache.spark.broadcast.TorrentBroadcastFactory.unbroadcast(TorrentBroadcastFactory.scala:45)
 at
 org.apache.spark.broadcast.BroadcastManager.unbroadcast(BroadcastManager.scala:66)
 at
 org.apache.spark.ContextCleaner.doCleanupBroadcast(ContextCleaner.scala:185)
 at
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:147)
 at
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:138)
 at scala.Option.foreach(Option.scala:236)
 at
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply$mcV$sp(ContextCleaner.scala:138)
 at
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
 at
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
 at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1460)
 at org.apache.spark.ContextCleaner.org
 http://org.apache.spark.contextcleaner.org/
 $apache$spark$ContextCleaner$$keepCleaning(ContextCleaner.scala:133)
 at org.apache.spark.ContextCleaner$$anon$3.run(ContextCleaner.scala:65)
 [Stage 91:===  (2 +
 4) / 6]15/01/29 18:08:15 ERROR SparkDeploySchedulerBackend: Asked to remove
 non-existent executor 0
 [Stage 93:  (29 + 20)
 / 49]15/01/29 23:47:03 ERROR TaskSchedulerImpl: Lost executor 9 on
 small11-tap1.common.lip6.fr: remote Akka client disassociated
 [Stage 83:   (1 + 0) / 6][Stage 86:   (0 + 1) / 2][Stage 88:   (0 + 2)
 / 8]15/01/29 23:47:06 ERROR SparkDeploySchedulerBackend: Asked to remove
 non-existent executor 9
 [Stage 83:===  (5 + 1) / 6][Stage 88:=   (9 + 2)
 / 11]15/01/29 23:57:30 ERROR TaskSchedulerImpl: Lost executor 8 on
 small10-tap1.common.lip6.fr: remote Akka client disassociated
 15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove
 non-existent executor 8
 15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove
 non-existent executor 8

 Best,
 Yifan LI









Re: [Graphx Spark] Error of Lost executor and TimeoutException

2015-01-30 Thread Yifan LI
Yes, I think so, esp. for a pregel application… have any suggestion?

Best,
Yifan LI





 On 30 Jan 2015, at 22:25, Sonal Goyal sonalgoy...@gmail.com wrote:
 
 Is your code hitting frequent garbage collection? 
 
 Best Regards,
 Sonal
 Founder, Nube Technologies http://www.nubetech.co/ 
 
  http://in.linkedin.com/in/sonalgoyal
 
 
 
 On Fri, Jan 30, 2015 at 7:52 PM, Yifan LI iamyifa...@gmail.com 
 mailto:iamyifa...@gmail.com wrote:
 
 
 
 Hi,
 
 I am running my graphx application on Spark 1.2.0(11 nodes cluster), has 
 requested 30GB memory per node and 100 cores for around 1GB input dataset(5 
 million vertices graph).
 
 But the error below always happen…
 
 Is there anyone could give me some points? 
 
 (BTW, the overall edge/vertex RDDs will reach more than 100GB during graph 
 computation, and another version of my application can work well on the same 
 dataset while it need much less memory during computation)
 
 Thanks in advance!!!
 
 
 15/01/29 18:05:08 ERROR ContextCleaner: Error cleaning broadcast 60
 java.util.concurrent.TimeoutException: Futures timed out after [30 seconds]
  at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
  at 
 scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
  at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
  at 
 scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
  at scala.concurrent.Await$.result(package.scala:107)
  at 
 org.apache.spark.storage.BlockManagerMaster.removeBroadcast(BlockManagerMaster.scala:137)
  at 
 org.apache.spark.broadcast.TorrentBroadcast$.unpersist(TorrentBroadcast.scala:227)
  at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.unbroadcast(TorrentBroadcastFactory.scala:45)
  at 
 org.apache.spark.broadcast.BroadcastManager.unbroadcast(BroadcastManager.scala:66)
  at 
 org.apache.spark.ContextCleaner.doCleanupBroadcast(ContextCleaner.scala:185)
  at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:147)
  at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:138)
  at scala.Option.foreach(Option.scala:236)
  at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply$mcV$sp(ContextCleaner.scala:138)
  at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
  at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
  at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1460)
  at org.apache.spark.ContextCleaner.org 
 http://org.apache.spark.contextcleaner.org/$apache$spark$ContextCleaner$$keepCleaning(ContextCleaner.scala:133)
  at org.apache.spark.ContextCleaner$$anon$3.run(ContextCleaner.scala:65)
 [Stage 91:===  (2 + 4) 
 / 6]15/01/29 18:08:15 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 0
 [Stage 93:  (29 + 20) / 
 49]15/01/29 23:47:03 ERROR TaskSchedulerImpl: Lost executor 9 on 
 small11-tap1.common.lip6.fr http://small11-tap1.common.lip6.fr/: remote 
 Akka client disassociated
 [Stage 83:   (1 + 0) / 6][Stage 86:   (0 + 1) / 2][Stage 88:   (0 + 2) / 
 8]15/01/29 23:47:06 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 9
 [Stage 83:===  (5 + 1) / 6][Stage 88:=   (9 + 2) / 
 11]15/01/29 23:57:30 ERROR TaskSchedulerImpl: Lost executor 8 on 
 small10-tap1.common.lip6.fr http://small10-tap1.common.lip6.fr/: remote 
 Akka client disassociated
 15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 8
 15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 8
 
 Best,
 Yifan LI
 
 
 
 
 
 
 



[Graphx Spark] Error of Lost executor and TimeoutException

2015-01-30 Thread Yifan LI

 
 
 Hi,
 
 I am running my graphx application on Spark 1.2.0(11 nodes cluster), has 
 requested 30GB memory per node and 100 cores for around 1GB input dataset(5 
 million vertices graph).
 
 But the error below always happen…
 
 Is there anyone could give me some points? 
 
 (BTW, the overall edge/vertex RDDs will reach more than 100GB during graph 
 computation, and another version of my application can work well on the same 
 dataset while it need much less memory during computation)
 
 Thanks in advance!!!
 
 
 15/01/29 18:05:08 ERROR ContextCleaner: Error cleaning broadcast 60
 java.util.concurrent.TimeoutException: Futures timed out after [30 seconds]
   at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
   at 
 scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
   at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
   at 
 scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
   at scala.concurrent.Await$.result(package.scala:107)
   at 
 org.apache.spark.storage.BlockManagerMaster.removeBroadcast(BlockManagerMaster.scala:137)
   at 
 org.apache.spark.broadcast.TorrentBroadcast$.unpersist(TorrentBroadcast.scala:227)
   at 
 org.apache.spark.broadcast.TorrentBroadcastFactory.unbroadcast(TorrentBroadcastFactory.scala:45)
   at 
 org.apache.spark.broadcast.BroadcastManager.unbroadcast(BroadcastManager.scala:66)
   at 
 org.apache.spark.ContextCleaner.doCleanupBroadcast(ContextCleaner.scala:185)
   at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:147)
   at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:138)
   at scala.Option.foreach(Option.scala:236)
   at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply$mcV$sp(ContextCleaner.scala:138)
   at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
   at 
 org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
   at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1460)
   at org.apache.spark.ContextCleaner.org 
 http://org.apache.spark.contextcleaner.org/$apache$spark$ContextCleaner$$keepCleaning(ContextCleaner.scala:133)
   at org.apache.spark.ContextCleaner$$anon$3.run(ContextCleaner.scala:65)
 [Stage 91:===  (2 + 4) / 
 6]15/01/29 18:08:15 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 0
 [Stage 93:  (29 + 20) / 
 49]15/01/29 23:47:03 ERROR TaskSchedulerImpl: Lost executor 9 on 
 small11-tap1.common.lip6.fr http://small11-tap1.common.lip6.fr/: remote 
 Akka client disassociated
 [Stage 83:   (1 + 0) / 6][Stage 86:   (0 + 1) / 2][Stage 88:   (0 + 2) / 
 8]15/01/29 23:47:06 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 9
 [Stage 83:===  (5 + 1) / 6][Stage 88:=   (9 + 2) / 
 11]15/01/29 23:57:30 ERROR TaskSchedulerImpl: Lost executor 8 on 
 small10-tap1.common.lip6.fr http://small10-tap1.common.lip6.fr/: remote 
 Akka client disassociated
 15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 8
 15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove 
 non-existent executor 8
 
 Best,
 Yifan LI
 
 
 
 
 



[Graphx Spark] Error of Lost executor and TimeoutException

2015-01-30 Thread Yifan LI
Hi,

I am running my graphx application on Spark 1.2.0(11 nodes cluster), has 
requested 30GB memory per node and 100 cores for around 1GB input dataset(5 
million vertices graph).

But the error below always happen…

Is there anyone could give me some points? 

(BTW, the overall edge/vertex RDDs will reach more than 100GB during graph 
computation, and another version of my application can work well on the same 
dataset while it need much less memory during computation)

Thanks in advance!!!


15/01/29 18:05:08 ERROR ContextCleaner: Error cleaning broadcast 60
java.util.concurrent.TimeoutException: Futures timed out after [30 seconds]
at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
at 
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
at scala.concurrent.Await$.result(package.scala:107)
at 
org.apache.spark.storage.BlockManagerMaster.removeBroadcast(BlockManagerMaster.scala:137)
at 
org.apache.spark.broadcast.TorrentBroadcast$.unpersist(TorrentBroadcast.scala:227)
at 
org.apache.spark.broadcast.TorrentBroadcastFactory.unbroadcast(TorrentBroadcastFactory.scala:45)
at 
org.apache.spark.broadcast.BroadcastManager.unbroadcast(BroadcastManager.scala:66)
at 
org.apache.spark.ContextCleaner.doCleanupBroadcast(ContextCleaner.scala:185)
at 
org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:147)
at 
org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1$$anonfun$apply$mcV$sp$2.apply(ContextCleaner.scala:138)
at scala.Option.foreach(Option.scala:236)
at 
org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply$mcV$sp(ContextCleaner.scala:138)
at 
org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
at 
org.apache.spark.ContextCleaner$$anonfun$org$apache$spark$ContextCleaner$$keepCleaning$1.apply(ContextCleaner.scala:134)
at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1460)
at 
org.apache.spark.ContextCleaner.org$apache$spark$ContextCleaner$$keepCleaning(ContextCleaner.scala:133)
at org.apache.spark.ContextCleaner$$anon$3.run(ContextCleaner.scala:65)
[Stage 91:===  (2 + 4) / 
6]15/01/29 18:08:15 ERROR SparkDeploySchedulerBackend: Asked to remove 
non-existent executor 0
[Stage 93:  (29 + 20) / 
49]15/01/29 23:47:03 ERROR TaskSchedulerImpl: Lost executor 9 on 
small11-tap1.common.lip6.fr: remote Akka client disassociated
[Stage 83:   (1 + 0) / 6][Stage 86:   (0 + 1) / 2][Stage 88:   (0 + 2) / 
8]15/01/29 23:47:06 ERROR SparkDeploySchedulerBackend: Asked to remove 
non-existent executor 9
[Stage 83:===  (5 + 1) / 6][Stage 88:=   (9 + 2) / 
11]15/01/29 23:57:30 ERROR TaskSchedulerImpl: Lost executor 8 on 
small10-tap1.common.lip6.fr: remote Akka client disassociated
15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove 
non-existent executor 8
15/01/29 23:57:30 ERROR SparkDeploySchedulerBackend: Asked to remove 
non-existent executor 8

Best,
Yifan LI