Github user gatorsmile commented on a diff in the pull request: https://github.com/apache/spark/pull/20610#discussion_r168347783 --- Diff: docs/sql-programming-guide.md --- @@ -1004,6 +1004,24 @@ Configuration of Parquet can be done using the `setConf` method on `SparkSession </tr> </table> +## ORC Files + +Since Spark 2.3, Spark supports a vectorized ORC reader with a new ORC file format for ORC files. To do that, the following configurations are newly added. The vectorized reader is used for the native ORC tables (e.g., the ones created using the clause `USING ORC`) when `spark.sql.orc.impl` is set to `native` and `spark.sql.orc.enableVectorizedReader` is set to `true`. For the Hive ORC serde table (e.g., the ones created using the clause `USING HIVE OPTIONS (fileFormat 'ORC')`), the vectorized reader is used when `spark.sql.hive.convertMetastoreOrc` is set to `true`. --- End diff -- `table ` -> `tables `
--- --------------------------------------------------------------------- To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org