[ 
https://issues.apache.org/jira/browse/SPARK-19301?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean Owen resolved SPARK-19301.
-------------------------------
    Resolution: Duplicate

> SparkContext is ignoring SparkConf when _jvm is not initialized on 
> spark-submit
> -------------------------------------------------------------------------------
>
>                 Key: SPARK-19301
>                 URL: https://issues.apache.org/jira/browse/SPARK-19301
>             Project: Spark
>          Issue Type: Bug
>          Components: PySpark
>    Affects Versions: 2.1.0
>            Reporter: Teppei Daito
>            Priority: Critical
>
> When using spark-submit with code below
> {code}
> SparkContext(conf=SparkConf().setAppName('foo'))
> {code}
> SparkContext ignores conf arg.
> This bug is started by this commit.
> https://github.com/apache/spark/commit/5b77e66dd6a128c5992ab3bde418613f84be7009
> *ignoring conf which not having _jconf*
> https://github.com/apache/spark/blob/5b77e66dd6a128c5992ab3bde418613f84be7009/python/pyspark/context.py#L125
> To resolve this problem, you have to call SparkContext._ensure_initialized() 
> before calling SparkConf().
> As I cannot find test code for the commit above, 
> SparkContext initialization process was too complicated for me to write a 
> patch to fix this problem.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to