[
https://issues.apache.org/jira/browse/TOREE-353?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15771997#comment-15771997
]
Harpreet Sawhney commented on TOREE-353:
----------------------------------------
I am facing similar issue with spark 1.6.3
> Spark defaulting to internal metastore instead of hive
> ------------------------------------------------------
>
> Key: TOREE-353
> URL: https://issues.apache.org/jira/browse/TOREE-353
> Project: TOREE
> Issue Type: Bug
> Affects Versions: 0.1.0
> Environment: Vanilla Hadoop 2.7.3, Hive 2.1.0, Spark 2.0.2 with
> Hadoop 2.7 binary distribution, Toree build from git repo and then installed
> using pip from torte-pip folder
> Reporter: Abhijit Bhole
>
> I cannot see Hive tables inside Toree kernel.
> For testing I have created a table in Hive and I execute the query "show
> tables from default" in spark. Some more details -
> - I have added the hive-submit.conf in SPARK_HOME/conf directory.
> - pyspark and spark-submit scripts work file in both yarn and local mode
> without any additional arguments. The pyspark shell also works fine and
> returns the table name.
> - On the driver UI page (on 4040 port) I can see
> spark.sql.catalogImplementation set to hive when running pyspark shell. But
> the variable is missing when running Toree kernel.
> - I have tried also giving --driver-class-path hive-site.xml --files
> /opt/hive-2.1.0/conf/hive-site.xml unsuccessfully
> I have tried everything I could think of hence filing a bug report. I have
> missed something please help.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)