[ 
https://issues.apache.org/jira/browse/SPARK-18750?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15729665#comment-15729665
 ] 

Sean Owen commented on SPARK-18750:
-----------------------------------

It sounds like you're focused on the StackOverflowError, and so controlling the 
number of executors is actually just incidental. There's not yet an indication 
here that the error comes from Spark though. Where is the stack trace coming 
from? If it's from YARN libraries the fix would have to go there.

> spark should be able to control the number of executor and should not throw 
> stack overslow
> ------------------------------------------------------------------------------------------
>
>                 Key: SPARK-18750
>                 URL: https://issues.apache.org/jira/browse/SPARK-18750
>             Project: Spark
>          Issue Type: Bug
>            Reporter: Neerja Khattar
>
> When running Sql queries on large datasets. Job fails with stack overflow 
> warning and it shows it is requesting lots of executors.
> Looks like there is no limit to number of executors or not even having an 
> upperbound based on yarn available resources.
> 16/11/29 15:47:47 INFO impl.ContainerManagementProtocolProxy: Opening proxy : 
> bdtcstr61n5.svr.us.jpmchase.net:8041 
> 16/11/29 15:47:47 INFO impl.ContainerManagementProtocolProxy: Opening proxy : 
> bdtcstr61n8.svr.us.jpmchase.net:8041 
> 16/11/29 15:47:47 INFO impl.ContainerManagementProtocolProxy: Opening proxy : 
> bdtcstr61n2.svr.us.jpmchase.net:8041 
> 16/11/29 15:47:47 INFO yarn.YarnAllocator: Driver requested a total number of 
> 32770 executor(s). 
> 16/11/29 15:47:47 INFO yarn.YarnAllocator: Will request 24576 executor 
> containers, each with 1 cores and 6758 MB memory including 614 MB overhead 
> 16/11/29 15:49:11 INFO yarn.YarnAllocator: Driver requested a total number of 
> 52902 executor(s). 
> 16/11/29 15:49:11 WARN yarn.ApplicationMaster: Reporter thread fails 1 
> time(s) in a row. 
> java.lang.StackOverflowError 
> at scala.collection.immutable.HashMap.$plus(HashMap.scala:57) 
> at scala.collection.immutable.HashMap.$plus(HashMap.scala:36) 
> at scala.collection.mutable.MapBuilder.$plus$eq(MapBuilder.scala:28) 
> If you notice in the error above, YARN is trying to request 24576 executor 
> containers, whereas the available cores are 1719. The Driver is requesting 
> for 52902 executor(s), which too high. 
> This exception should be fixed



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to