Github user srowen commented on a diff in the pull request:

    https://github.com/apache/spark/pull/22453#discussion_r219722694
  
    --- Diff: docs/sql-programming-guide.md ---
    @@ -1002,6 +1002,15 @@ Configuration of Parquet can be done using the 
`setConf` method on `SparkSession
         </p>
       </td>
     </tr>
    +<tr>
    +  <td><code>spark.sql.parquet.writeLegacyFormat</code></td>
    --- End diff --
    
    OK that sounds important to document. But the reasoning in this thread is 
also more useful information I think. Instead of describing it as a legacy 
format (implying it's not valid Parquet or something) and that it's required 
for Hive and Impala, can we mention or point to the specific reason that would 
cause you to need this? The value of the documentation here is in whether it 
helps the user know when to set it one way or the other.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to