GitHub user liancheng opened a pull request:
https://github.com/apache/spark/pull/5348
[Doc] [SQL] Addes Hive metastore Parquet table conversion section
This PR adds a section about Hive metastore Parquet table conversion. It
documents:
1. Schema reconciliation rules introduced in #5214 (see [this comment] [1]
in #5188)
2. Metadata refreshing requirement introduced in #5339
Notice that Python snippet for refreshing tables is not included, because
we don't have `refreshTable` in PySpark. This should be addressed in a separate
PR.
[1]: https://github.com/apache/spark/pull/5188#issuecomment-86531248
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/liancheng/spark sql-doc-parquet-conversion
Alternatively you can review and apply these changes as the patch at:
https://github.com/apache/spark/pull/5348.patch
To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:
This closes #5348
----
commit 9840affeb34c2c7ee21cec366af36bc655c4b4fc
Author: Cheng Lian <[email protected]>
Date: 2015-04-03T12:56:53Z
Addes Hive metastore Parquet table conversion section
----
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]