Hi,

I am getting the below error and due to this there is no completed stages-
all the waiting

*14/12/05 03:31:59 WARN AkkaUtils: Error sending message in 1 attempts*
*java.util.concurrent.TimeoutException: Futures timed out after [30
seconds]*
*        at
scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)*
*        at
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)*
*        at
scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)*
*        at
akka.dispatch.MonitorableThreadFactory$AkkaForkJoinWorkerThread$$anon$3.block(ThreadPoolBuilder.scala:169)*
*        at
scala.concurrent.forkjoin.ForkJoinPool.managedBlock(ForkJoinPool.java:3640)*
*        at
akka.dispatch.MonitorableThreadFactory$AkkaForkJoinWorkerThread.blockOn(ThreadPoolBuilder.scala:167)*
*        at scala.concurrent.Await$.result(package.scala:107)*
*        at
org.apache.spark.util.AkkaUtils$.askWithReply(AkkaUtils.scala:176)*
*        at
org.apache.spark.storage.BlockManagerMaster.askDriverWithReply(BlockManagerMaster.scala:213)*
*        at
org.apache.spark.storage.BlockManagerMaster.tell(BlockManagerMaster.scala:203)*
*        at
org.apache.spark.storage.BlockManagerMaster.registerBlockManager(BlockManagerMaster.scala:47)*
*        at
org.apache.spark.storage.BlockManager.initialize(BlockManager.scala:177)*
*        at
org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:147)*
*        at
org.apache.spark.storage.BlockManager.<init>(BlockManager.scala:168)*
*        at org.apache.spark.SparkEnv$.create(SparkEnv.scala:230)*
*        at org.apache.spark.executor.Executor.<init>(Executor.scala:78)*
*        at
org.apache.spark.executor.CoarseGrainedExecutorBackend$$anonfun$receiveWithLogging$1.applyOrElse(CoarseGrainedExecutorBackend.scala:60)*
*        at
scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)*
*        at
scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)*
*        at
scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25)*
*        at
org.apache.spark.util.ActorLogReceive$$anon$1.apply(ActorLogReceive.scala:53)*
*        at
org.apache.spark.util.ActorLogReceive$$anon$1.apply(ActorLogReceive.scala:42)*
*        at
scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118)*
*        at
org.apache.spark.util.ActorLogReceive$$anon$1.applyOrElse(ActorLogReceive.scala:42)*
*        at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)*
*        at akka.actor.ActorCell.invoke(ActorCell.scala:456)*
*        at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)*
*        at akka.dispatch.Mailbox.run(Mailbox.scala:219)*
*        at
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)*
*        at
scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)*
*        at
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)*
*        at
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)*
*        at
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)*


Could you please let me know the reason and fix for this? Spark version is
1.1.1

-- 

Sourav Chandra

Senior Software Engineer

· · · · · · · · · · · · · · · · · · · · · · · · · · · · · · · · ·

sourav.chan...@livestream.com

o: +91 80 4121 8723

m: +91 988 699 3746

skype: sourav.chandra

Livestream

"Ajmera Summit", First Floor, #3/D, 68 Ward, 3rd Cross, 7th C Main, 3rd
Block, Koramangala Industrial Area,

Bangalore 560034

www.livestream.com

Reply via email to