viirya commented on a change in pull request #28366:
URL: https://github.com/apache/spark/pull/28366#discussion_r420615817
##########
File path:
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala
##########
@@ -2063,16 +2063,17 @@ object SQLConf {
.booleanConf
.createWithDefault(true)
- val NESTED_PREDICATE_PUSHDOWN_ENABLED =
- buildConf("spark.sql.optimizer.nestedPredicatePushdown.enabled")
- .internal()
- .doc("When true, Spark tries to push down predicates for nested columns
and or names " +
- "containing `dots` to data sources. Currently, Parquet implements both
optimizations " +
- "while ORC only supports predicates for names containing `dots`. The
other data sources" +
- "don't support this feature yet.")
+ val NESTED_PREDICATE_PUSHDOWN_FILE_SOURCE_LIST =
+
buildConf("spark.sql.optimizer.nestedPredicatePushdown.supportedFileSources")
+ .internal()
+ .doc("A comma-separated list of data source short names or fully
qualified data source " +
+ "implementation class names for which Spark tries to push down
predicates for nested " +
+ "columns and/or names containing `dots` to data sources. Currently,
Parquet implements " +
+ "both optimizations while ORC only supports predicates for names
containing `dots`. The " +
+ "other data sources don't support this feature yet. So the default
value is 'parquet,orc'.")
Review comment:
I think DSv2 API supposes nested column capacity like pushdown and
pruning, so we only need to deal with DSv1 compatibility issues here.
Precisely, file source.
I will create a simple followup to refine the doc of this configuration for
this point. Thanks.
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]