[
https://issues.apache.org/jira/browse/SPARK-19301?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Sean Owen resolved SPARK-19301.
-------------------------------
Resolution: Duplicate
> SparkContext is ignoring SparkConf when _jvm is not initialized on
> spark-submit
> -------------------------------------------------------------------------------
>
> Key: SPARK-19301
> URL: https://issues.apache.org/jira/browse/SPARK-19301
> Project: Spark
> Issue Type: Bug
> Components: PySpark
> Affects Versions: 2.1.0
> Reporter: Teppei Daito
> Priority: Critical
>
> When using spark-submit with code below
> {code}
> SparkContext(conf=SparkConf().setAppName('foo'))
> {code}
> SparkContext ignores conf arg.
> This bug is started by this commit.
> https://github.com/apache/spark/commit/5b77e66dd6a128c5992ab3bde418613f84be7009
> *ignoring conf which not having _jconf*
> https://github.com/apache/spark/blob/5b77e66dd6a128c5992ab3bde418613f84be7009/python/pyspark/context.py#L125
> To resolve this problem, you have to call SparkContext._ensure_initialized()
> before calling SparkConf().
> As I cannot find test code for the commit above,
> SparkContext initialization process was too complicated for me to write a
> patch to fix this problem.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]