Thanks for your reply, I have tried your suggestion. it works now.

At 2016-07-28 16:18:01, "Sun Rui" <sunrise_...@163.com> wrote:
Yes, this is a change in Spark 2.0.  you can take a look at 
https://issues.apache.org/jira/browse/SPARK-13723


In the latest Spark On Yarn documentation for Spark 2.0, there is updated 
description for --num-executors:
| spark.executor.instances | 2 | The number of executors for static allocation. 
Withspark.dynamicAllocation.enabled, the initial set of executors will be at 
least this large. |
You can disable the dynamic allocation for an application by specifying “--conf 
spark.dynamicAllocation.enabled=false” in the command line.


On Jul 28, 2016, at 15:44, LONG WANG <wanglong_...@163.com> wrote:


Hi Spark Experts,


                          Today I tried Spark 2.0 on YARN and also enabled 
Dynamic Resource Allocation feature, I just find that no matter I specify 
--num-executor in spark-submit command or not, the Dynamic Resource Allocation 
is used, but I remember when I specify --num-executor option in spark-submit 
command in Spark 1.6, the Dynamic Resource Allocation feature will not be 
used/effect for that job. And I can see below log in Spark 1.6 .


                            <截图1.png>
                           
                          Is this a behavior change in Spark 2.0? And How can I 
disable Dynamic Resource Allocation for a specific job submission temporarily 
as before? 



 

 邮件带有附件预览链接,若您转发或回复此邮件时不希望对方预览附件,建议您手动删除链接。
共有 1 个附件
截图1.png(23K)极速下载 在线预览

Reply via email to