[ 
https://issues.apache.org/jira/browse/SPARK-7108?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14514836#comment-14514836
 ] 

Marcelo Vanzin commented on SPARK-7108:
---------------------------------------

bq. It might be that we just document as such and move on

But that's my point. The documentation already unequivocally states that {{this 
will be overriden}}. It doesn't say "might be overridden if the cluster manager 
is configured a certain way". In YARN, that's the way it has worked since 1.0. 
Standalone works differently because it, intentionally or not, supports an 
undocumented way of setting up that config in executors.

> spark.local.dir is no longer honored in Standalone mode
> -------------------------------------------------------
>
>                 Key: SPARK-7108
>                 URL: https://issues.apache.org/jira/browse/SPARK-7108
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>    Affects Versions: 1.2.1, 1.3.0
>            Reporter: Josh Rosen
>            Priority: Critical
>
> Prior to SPARK-4834, configuring spark.local.dir in the driver would affect 
> the local directories created on the executor.  After this patch, executors 
> will always ignore this setting in favor of directories read from 
> {{SPARK_LOCAL_DIRS}}, which is set by the standalone worker based on the 
> worker's own configuration and not the application configuration.
> This change impacts users who configured {{spark.local.dir}} only in their 
> driver and not via their cluster's {{spark-defaults.conf}} or 
> {{spark-env.sh}} files.  This is an atypical use-case, since the available 
> local directories / disks are a property of the cluster and not the 
> application, which probably explains why this issue has not been reported 
> previously.
> The correct fix might be comment + documentation improvements.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to