Github user srowen commented on a diff in the pull request:
https://github.com/apache/spark/pull/22453#discussion_r219722694
--- Diff: docs/sql-programming-guide.md ---
@@ -1002,6 +1002,15 @@ Configuration of Parquet can be done using the
`setConf` method on `SparkSession
</p>
</td>
</tr>
+<tr>
+ <td><code>spark.sql.parquet.writeLegacyFormat</code></td>
--- End diff --
OK that sounds important to document. But the reasoning in this thread is
also more useful information I think. Instead of describing it as a legacy
format (implying it's not valid Parquet or something) and that it's required
for Hive and Impala, can we mention or point to the specific reason that would
cause you to need this? The value of the documentation here is in whether it
helps the user know when to set it one way or the other.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]