Github user sryza commented on the pull request:

    https://github.com/apache/spark/pull/4051#issuecomment-71406518
  
    > @sryza Is the point of not requiring these configs that the users don't 
really know how many executors they actually want?
    
    Exactly.  From my perspective, one of the main motivations for dynamic 
allocation in the first place was removing the need to set `--num-executors`.  
People have long complained about the long list configurations users need to 
know for MapReduce, but at this point Spark actually requires more to get 
decent resource utilization.  This particular JIRA was filed Hive didn't know 
how to set the max executors in a general way.  What specific advice would you 
give to a user about how to set this property?
    
    > Also defaulting the max to Int.MaxValue doesn't make much sense in my 
opinion because the whole point of doing this is to share resources in the 
cluster.
    
    This is the job of the cluster resource manager.  Users and apps shouldn't 
need self-regulate their usage. YARN will throttle resources when there is 
contention.  The main responsibility of the app is just to not ask for / sit on 
more resources than it can take advantage of, which dynamic allocation helps 
with in general, and which SPARK-4136 will help enforce more strictly.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to