The configuration key should be "spark.akka.askTimeout" for this timeout.
The time unit is seconds.

Best Regards,
Shixiong(Ryan) Zhu

2015-04-26 15:15 GMT-07:00 Deepak Gopalakrishnan <dgk...@gmail.com>:

> Hello,
>
>
> Just to add a bit more context :
>
> I have done that in the code, but I cannot see it change from 30 seconds
> in the log.
>
>                 .set("spark.executor.memory", "10g")
>
>                 .set("spark.driver.memory", "20g")
>
>                 .set("spark.akka.timeout","6000")
>
> PS : I understand that 6000 is quite large, but I'm just trying to see if
> it actually changes
>
>
> Here is the command that I'm running
>
>  sudo MASTER=spark://master.spark.com:7077
> /opt/spark/spark-1.3.0-bin-hadoop2.4/bin/spark-submit --class
> "<class-name>"   --executor-memory 20G --driver-memory 10G  --deploy-mode
> client --conf spark.akka.timeout=6000 --conf spark.akka.askTimeout=6000
> <jar file path>
>
>
> and here is how I load the file         JavaPairRDD<String, String>
> learningRdd=sc.wholeTextFiles(filePath,10);
> Thanks
>
> On Mon, Apr 27, 2015 at 3:36 AM, Bryan Cutler <cutl...@gmail.com> wrote:
>
>> I'm not sure what the expected performance should be for this amount of
>> data, but you could try to increase the timeout with the property
>> "spark.akka.timeout" to see if that helps.
>>
>> Bryan
>>
>> On Sun, Apr 26, 2015 at 6:57 AM, Deepak Gopalakrishnan <dgk...@gmail.com>
>> wrote:
>>
>>> Hello All,
>>>
>>> I'm trying to process a 3.5GB file on standalone mode using spark. I
>>> could run my spark job succesfully on a 100MB file and it works as
>>> expected. But, when I try to run it on the 3.5GB file, I run into the below
>>> error :
>>>
>>>
>>> 15/04/26 12:45:50 INFO BlockManagerMaster: Updated info of block 
>>> taskresult_83
>>> 15/04/26 12:46:46 WARN AkkaUtils: Error sending message [message = 
>>> Heartbeat(2,[Lscala.Tuple2;@790223d3,BlockManagerId(2, master.spark.com, 
>>> 39143))] in 1 attempts
>>> java.util.concurrent.TimeoutException: Futures timed out after [30 seconds]
>>>     at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>>     at 
>>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>>>     at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>>     at 
>>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>>>     at scala.concurrent.Await$.result(package.scala:107)
>>>     at org.apache.spark.util.AkkaUtils$.askWithReply(AkkaUtils.scala:195)
>>>     at org.apache.spark.executor.Executor$$anon$1.run(Executor.scala:427)
>>> 15/04/26 12:47:15 INFO MemoryStore: ensureFreeSpace(26227673) called with 
>>> curMem=265897, maxMem=5556991426
>>> 15/04/26 12:47:15 INFO MemoryStore: Block taskresult_92 stored as bytes in 
>>> memory (estimated size 25.0 MB, free 5.2 GB)
>>> 15/04/26 12:47:16 INFO MemoryStore: ensureFreeSpace(26272879) called with 
>>> curMem=26493570, maxMem=5556991426
>>> 15/04/26 12:47:16 INFO MemoryStore: Block taskresult_94 stored as bytes in 
>>> memory (estimated size 25.1 MB, free 5.1 GB)
>>> 15/04/26 12:47:18 INFO MemoryStore: ensureFreeSpace(26285327) called with 
>>> curMem=52766449, maxMem=5556991426
>>>
>>>
>>> and the job fails.
>>>
>>>
>>> I'm on AWS and have opened all ports. Also, since the 100MB file works,
>>> it should not be a connection issue.  I've a r3 xlarge and 2 m3 large.
>>>
>>> Can anyone suggest a way to fix this?
>>>
>>> --
>>> Regards,
>>> *Deepak Gopalakrishnan*
>>> *Mobile*:+918891509774
>>> *Skype* : deepakgk87
>>> http://myexps.blogspot.com
>>>
>>>
>>
>
>
> --
> Regards,
> *Deepak Gopalakrishnan*
> *Mobile*:+918891509774
> *Skype* : deepakgk87
> http://myexps.blogspot.com
>
>

Reply via email to