[ 
https://issues.apache.org/jira/browse/TOREE-353?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15774000#comment-15774000
 ] 

Marius Van Niekerk commented on TOREE-353:
------------------------------------------

Hive 2.1.0 may be causing some issues.  Spark by default uses hive 1.2 for the 
metastore.  Just passing the classpath is insufficient as you need to set the 
spark.sql.hive.metastore.jars property to pick these up.

Can you close this and ask your questions on the Spark mailing list as this is 
not part of Toree.

> Spark defaulting to internal metastore instead of hive
> ------------------------------------------------------
>
>                 Key: TOREE-353
>                 URL: https://issues.apache.org/jira/browse/TOREE-353
>             Project: TOREE
>          Issue Type: Bug
>    Affects Versions: 0.1.0
>         Environment: Vanilla Hadoop 2.7.3, Hive 2.1.0, Spark 2.0.2 with 
> Hadoop 2.7 binary distribution, Toree build from git repo and then installed 
> using pip from torte-pip folder
>            Reporter: Abhijit Bhole
>
> I cannot see Hive tables inside Toree kernel. 
> For testing I have created a table in Hive and I execute the query "show 
> tables from default" in spark. Some more details -
> - I have added the hive-submit.conf in SPARK_HOME/conf directory. 
> - pyspark and spark-submit scripts work file in both yarn and local mode 
> without any additional arguments. The pyspark shell also works fine and 
> returns the table name. 
> - On the driver UI page (on 4040 port) I can see  
> spark.sql.catalogImplementation set to hive when running pyspark shell. But 
> the variable is missing when running Toree kernel.
> - I have tried also giving --driver-class-path hive-site.xml --files 
> /opt/hive-2.1.0/conf/hive-site.xml unsuccessfully
> I have tried everything I could think of hence filing a bug report. I have 
> missed something please help.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to