[ 
https://issues.apache.org/jira/browse/SPARK-19664?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Song Jun updated SPARK-19664:
-----------------------------
    Description: 
In SPARK-15959, we bring back the 'hive.metastore.warehouse.dir' , while in the 
logic, when use the value of  'spark.sql.warehouse.dir' to overwrite 
'spark.sql.warehouse.dir' , it set it to 'sparkContext.conf' , I think it 
should put in 'sparkContext.hadoopConfiguration' and overwrite the original 
value of hadoopConf

https://github.com/apache/spark/blob/master/sql/core/src/main/scala/org/apache/spark/sql/internal/SharedState.scala#L64

  was:
In SPARK-15959, we bring back the 'hive.metastore.warehouse.dir' , while in the 
logic, when use the value of  'spark.sql.warehouse.dir' to overwrite 
'spark.sql.warehouse.dir' , it set it to 'sparkContext.conf' , I think it 
should put in 'sparkContext.hadoopConfiguration'

https://github.com/apache/spark/blob/master/sql/core/src/main/scala/org/apache/spark/sql/internal/SharedState.scala#L64


> put 'hive.metastore.warehouse.dir' in hadoopConf place
> ------------------------------------------------------
>
>                 Key: SPARK-19664
>                 URL: https://issues.apache.org/jira/browse/SPARK-19664
>             Project: Spark
>          Issue Type: Improvement
>          Components: SQL
>    Affects Versions: 2.1.0
>            Reporter: Song Jun
>            Priority: Minor
>
> In SPARK-15959, we bring back the 'hive.metastore.warehouse.dir' , while in 
> the logic, when use the value of  'spark.sql.warehouse.dir' to overwrite 
> 'spark.sql.warehouse.dir' , it set it to 'sparkContext.conf' , I think it 
> should put in 'sparkContext.hadoopConfiguration' and overwrite the original 
> value of hadoopConf
> https://github.com/apache/spark/blob/master/sql/core/src/main/scala/org/apache/spark/sql/internal/SharedState.scala#L64



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to