You can change "spark.sql.broadcastTimeout" to increase the timeout. The
default value is 300 seconds.

Best Regards,
Shixiong Zhu

2015-09-24 15:16 GMT+08:00 Eyad Sibai <eyad.alsi...@gmail.com>:

> I am trying to join two tables using dataframes using python 3.4 and I am
> getting the following error
>
>
> I ran it on my localhost machine with 2 workers, spark 1.5
>
>
> I always get timeout if the job takes more than 5 minutes.
>
>
>
> at
> org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:49)
>
>  at
> org.apache.spark.sql.execution.aggregate.TungstenAggregate.doExecute(TungstenAggregate.scala:69)
>
>  at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:140)
>
>  at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:138)
>
>  at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
>
>  at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:138)
>
>  at
> org.apache.spark.sql.execution.Exchange$$anonfun$doExecute$1.apply(Exchange.scala:142)
>
>  at
> org.apache.spark.sql.execution.Exchange$$anonfun$doExecute$1.apply(Exchange.scala:141)
>
>  at
> org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:48)
>
>  ... 33 more
>
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after
> [300 seconds]
>
>  at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>
>  at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>
>  at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>
>  at
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>
>  at scala.concurrent.Await$.result(package.scala:107)
>
>  at
> org.apache.spark.sql.execution.joins.BroadcastHashJoin.doExecute(BroadcastHashJoin.scala:110)
>
>  at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:140)
>
>  at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:138)
>
>  at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
>
>  at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:138)
>
>  at
> org.apache.spark.sql.execution.TungstenProject.doExecute(basicOperators.scala:86)
>
>  at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:140)
>
>  at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:138)
>
>  at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
>
>  at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:138)
>
>  at
> org.apache.spark.sql.execution.aggregate.TungstenAggregate$$anonfun$doExecute$1.apply(TungstenAggregate.scala:119)
>
>  at
> org.apache.spark.sql.execution.aggregate.TungstenAggregate$$anonfun$doExecute$1.apply(TungstenAggregate.scala:69)
>
>  at
> org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:48)
>
>  ... 41 more
>
>
> 2015-09-23 15:44:09,536 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/static/sql,null}
>
> 2015-09-23 15:44:09,537 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/SQL1/execution/json,null}
>
> 2015-09-23 15:44:09,538 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/SQL1/execution,null}
>
> 2015-09-23 15:44:09,539 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/SQL1/json,null}
>
> 2015-09-23 15:44:09,539 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/SQL1,null}
>
> 2015-09-23 15:44:09,539 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/static/sql,null}
>
> 2015-09-23 15:44:09,539 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/SQL/execution/json,null}
>
> 2015-09-23 15:44:09,539 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/SQL/execution,null}
>
> 2015-09-23 15:44:09,539 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/SQL/json,null}
>
> 2015-09-23 15:44:09,539 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/SQL,null}
>
> 2015-09-23 15:44:09,539 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/metrics/json,null}
>
> 2015-09-23 15:44:09,540 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/stages/stage/kill,null}
>
> 2015-09-23 15:44:09,540 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/api,null}
>
> 2015-09-23 15:44:09,540 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/,null}
>
> 2015-09-23 15:44:09,540 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/static,null}
>
> 2015-09-23 15:44:09,540 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/executors/threadDump/json,null}
>
> 2015-09-23 15:44:09,540 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/executors/threadDump,null}
>
> 2015-09-23 15:44:09,540 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/executors/json,null}
>
> 2015-09-23 15:44:09,540 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/executors,null}
>
> 2015-09-23 15:44:09,540 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/environment/json,null}
>
> 2015-09-23 15:44:09,540 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/environment,null}
>
> 2015-09-23 15:44:09,540 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/storage/rdd/json,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/storage/rdd,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/storage/json,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/storage,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/stages/pool/json,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/stages/pool,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/stages/stage/json,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/stages/stage,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/stages/json,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/stages,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/jobs/job/json,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/jobs/job,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/jobs/json,null}
>
> 2015-09-23 15:44:09,541 INFO ContextHandler: stopped
> o.s.j.s.ServletContextHandler{/jobs,null}
>
> 2015-09-23 15:44:09,595 ERROR DAGScheduler: Failed to update accumulators
> for ResultTask(0, 177)
>
> org.apache.spark.SparkException: EOF reached before Python server
> acknowledged
>
>  at
> org.apache.spark.api.python.PythonAccumulatorParam.addInPlace(PythonRDD.scala:830)
>
>  at
> org.apache.spark.api.python.PythonAccumulatorParam.addInPlace(PythonRDD.scala:788)
>
>  at org.apache.spark.Accumulable.$plus$plus$eq(Accumulators.scala:100)
>
>  at
> org.apache.spark.Accumulators$$anonfun$add$2.apply(Accumulators.scala:351)
>
>  at
> org.apache.spark.Accumulators$$anonfun$add$2.apply(Accumulators.scala:346)
>
>  at
> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772)
>
>  at
> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98)
>
>  at
> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98)
>
>  at
> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226)
>
>  at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39)
>
>  at scala.collection.mutable.HashMap.foreach(HashMap.scala:98)
>
>  at
> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771)
>
>  at org.apache.spark.Accumulators$.add(Accumulators.scala:346)
>
>  at
> org.apache.spark.scheduler.DAGScheduler.updateAccumulators(DAGScheduler.scala:934)
>
>  at
> org.apache.spark.scheduler.DAGScheduler.handleTaskCompletion(DAGScheduler.scala:1003)
>
>  at
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1490)
>
>  at
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1455)
>
>  at
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1444)
>
>  at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
>
> 2015-09-23 15:44:09,689 ERROR Executor: Exception in task 178.0 in stage
> 0.0 (TID 178)
>
> java.net.SocketException: Broken pipe
>
>  at java.net.SocketOutputStream.socketWrite0(Native Method)
>
>  at java.net.SocketOutputStream.socketWrite(SocketOutputStream.java:109)
>
>  at java.net.SocketOutputStream.write(SocketOutputStream.java:153)
>
>  at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
>
>  at java.io.BufferedOutputStream.write(BufferedOutputStream.java:126)
>
>  at java.io.DataOutputStream.write(DataOutputStream.java:107)
>
>  at java.io.FilterOutputStream.write(FilterOutputStream.java:97)
>
>  at org.apache.spark.api.python.PythonRDD$.writeUTF(PythonRDD.scala:590)
>
>  at
> org.apache.spark.api.python.PythonRDD$.org$apache$spark$api$python$PythonRDD$$write$1(PythonRDD.scala:410)
>
>  at
> org.apache.spark.api.python.PythonRDD$$anonfun$writeIteratorToStream$1.apply(PythonRDD.scala:420)
>
>  at
> org.apache.spark.api.python.PythonRDD$$anonfun$writeIteratorToStream$1.apply(PythonRDD.scala:420)
>
>  at scala.collection.Iterator$class.foreach(Iterator.scala:727)
>
>  at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
>
>  at
> org.apache.spark.api.python.PythonRDD$.writeIteratorToStream(PythonRDD.scala:420)
>
>  at
> org.apache.spark.api.python.PythonRDD$WriterThread$$anonfun$run$3.apply(PythonRDD.scala:249)
>
>  at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1699)
>
>  at
> org.apache.spark.api.python.PythonRDD$WriterThread.run(PythonRDD.scala:208)
>
> EYAD SIBAI
> Risk Engineer
>
> *iZettle ®*
> ––
>
> Mobile: +46 72 911 60 54 <+46%2072%20911%2060%2054>
> Web: www.izettle.com <http://izettle.com/>
>

Reply via email to