[ 
https://issues.apache.org/jira/browse/AIRFLOW-6086?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Florian FERREIRA closed AIRFLOW-6086.
-------------------------------------
    Resolution: Duplicate

> SparkSubmitOperator - Unable to override spark_binary 
> ------------------------------------------------------
>
>                 Key: AIRFLOW-6086
>                 URL: https://issues.apache.org/jira/browse/AIRFLOW-6086
>             Project: Apache Airflow
>          Issue Type: Bug
>          Components: contrib, core
>    Affects Versions: 1.10.6
>            Reporter: Florian FERREIRA
>            Priority: Major
>
> Hello,
> I have a connection "spark2_default" : 
> || Conn Id || Conn Type ||  Host || Port || Is Encrypted || Is Extra 
> Encrypted || Extra  
> | 'spark2_default' | 'spark2'  | 'yarn-cluster' | None | False | False | 
> {"master":"yarn-cluster","deploy-mode":"cluster","spark-binary":"spark2-submit"}
>  |
> Extra contains 'spark-binary' key that was use by airflow 1.10.2 to choose 
> spark-submit operator. But in version 1.10.6 this config is ignore.
> I think that , in class SparkSubmitOperator in init function they has a 
> default value "spark-submit" for spark_binary parameter. 
> {code}
>  spark_binary="spark-submit",
> {code}
> Therefore in class SparkSubmitHook when we control if spark_binary is empty 
> it can't be.
> {code}
> conn_data['spark_binary'] = self._spark_binary or  \
>                 extra.get('spark-binary', "spark-submit")
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to