gatorsmile commented on a change in pull request #24518: [SPARK-27627][SQL] 
Make option "pathGlobFilter" as a general option for all file sources
URL: https://github.com/apache/spark/pull/24518#discussion_r280829502
 
 

 ##########
 File path: docs/sql-migration-guide-upgrade.md
 ##########
 @@ -126,6 +126,8 @@ license: |
 
   - Since Spark 3.0, parquet logical type `TIMESTAMP_MICROS` is used by 
default while saving `TIMESTAMP` columns. In Spark version 2.4 and earlier, 
`TIMESTAMP` columns are saved as `INT96` in parquet files. To set `INT96` to 
`spark.sql.parquet.outputTimestampType` restores the previous behavior.
 
+  - Since Spark 3.0, a new data source option `pathGlobFilter` is introduced 
for filtering files in `DataFrameReader` and `DataStreamReader`. For example, 
`spark.read.option("pathGlobFilter", "*.orc").orc(path)` will read all the 
files ending with `.orc` under the given `path`. Note that with the option the 
query result will contain partition columns if any; while with glob pattern in 
path, e.g `spark.read.orc("path/*/*/*/*.orc")`, the result won't contain 
partition columns.
 
 Review comment:
   This does not introduce any behavior change. Thus, I would prefer to 
document it in the SQL guide: 
https://spark.apache.org/docs/latest/sql-data-sources-load-save-functions.html 
   
   I think we can add a section for documenting the useful options for our 
built-in file sources. 

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to