GitHub user liancheng opened a pull request:

    https://github.com/apache/spark/pull/5348

    [Doc] [SQL] Addes Hive metastore Parquet table conversion section

    This PR adds a section about Hive metastore Parquet table conversion. It 
documents:
    
    1. Schema reconciliation rules introduced in #5214 (see [this comment] [1] 
in #5188)
    2. Metadata refreshing requirement introduced in #5339
    
    Notice that Python snippet for refreshing tables is not included, because 
we don't have `refreshTable` in PySpark. This should be addressed in a separate 
PR.
    
    [1]: https://github.com/apache/spark/pull/5188#issuecomment-86531248

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/liancheng/spark sql-doc-parquet-conversion

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/spark/pull/5348.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #5348
    
----
commit 9840affeb34c2c7ee21cec366af36bc655c4b4fc
Author: Cheng Lian <[email protected]>
Date:   2015-04-03T12:56:53Z

    Addes Hive metastore Parquet table conversion section

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to