Github user liancheng commented on the pull request:

    https://github.com/apache/spark/pull/5733#issuecomment-101358044
  
    Here are my two cents:
    
    1. Short term: we simply don't allow data source tables persisted in Hive 
metastore to be accessible from Hive client API.
    2. Mid term: maybe a Hive storage handler similar to the one for HBase, 
responsible for translating Spark SQL data source table properties stored as 
Hive metastore SerDe properties to corresponding Hive concepts.
    3. Long term:
    
       - Migrate Spark SQL Hive support to a separate external data source
       - Have Spark SQL's own metastore service
       - The storage handler mentioned above can be used to interacting with 
this metastore service
    
       In a word, ideally, Hive is an external data source of Spark SQL via 
Spark SQL's external data sources API, while Spark SQL can also be viewed as a 
data source of Hive via Hive's storage handler API.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to