dongjoon-hyun commented on code in PR #36304:
URL: https://github.com/apache/spark/pull/36304#discussion_r989616936


##########
sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala:
##########
@@ -412,6 +412,21 @@ object SQLConf {
       .longConf
       .createWithDefault(67108864L)
 
+  val RUNTIME_ROW_LEVEL_OPERATION_GROUP_FILTER_ENABLED =
+    
buildConf("spark.sql.optimizer.runtime.rowLevelOperationGroupFilter.enabled")
+      .doc("Enables runtime group filtering for group-based row-level 
operations. " +
+        "Data sources that replace groups of data (e.g. files, partitions) may 
prune entire " +
+        "groups using provided data source filters when planning a row-level 
operation scan. " +
+        "However, such filtering is limited as not all expressions can be 
converted into data " +
+        "source filters and some expressions can only be evaluated by Spark 
(e.g. subqueries). " +
+        "Since rewriting groups is expensive, Spark can execute a query at 
runtime to find what " +
+        "records match the condition of the row-level operation. The 
information about matching " +
+        "records will be passed back to the row-level operation scan, allowing 
data sources to " +
+        "discard groups that don't have to be rewritten.")
+      .version("3.4.0")
+      .booleanConf
+      .createWithDefault(true)

Review Comment:
   I agree with starting with `true` in this case.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to