This is an automated email from the ASF dual-hosted git repository.

danny0405 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/master by this push:
     new 98b3d3bac0f [HUDI-7980] Optimize the configuration content when 
performing clustering with row writer (#11614)
98b3d3bac0f is described below

commit 98b3d3bac0f31219e5b93b7528516b27b87ea699
Author: majian <[email protected]>
AuthorDate: Sun Jul 14 09:06:37 2024 +0800

    [HUDI-7980] Optimize the configuration content when performing clustering 
with row writer (#11614)
---
 .../run/strategy/MultipleSparkJobExecutionStrategy.java           | 8 +++++---
 1 file changed, 5 insertions(+), 3 deletions(-)

diff --git 
a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/clustering/run/strategy/MultipleSparkJobExecutionStrategy.java
 
b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/clustering/run/strategy/MultipleSparkJobExecutionStrategy.java
index 75b42491eda..47ccd8700a8 100644
--- 
a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/clustering/run/strategy/MultipleSparkJobExecutionStrategy.java
+++ 
b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/clustering/run/strategy/MultipleSparkJobExecutionStrategy.java
@@ -93,7 +93,6 @@ import java.util.stream.Collectors;
 import java.util.stream.Stream;
 
 import static 
org.apache.hudi.client.utils.SparkPartitionUtils.getPartitionFieldVals;
-import static org.apache.hudi.common.config.HoodieCommonConfig.TIMESTAMP_AS_OF;
 import static 
org.apache.hudi.config.HoodieClusteringConfig.PLAN_STRATEGY_SORT_COLUMNS;
 import static 
org.apache.hudi.io.storage.HoodieSparkIOFactory.getHoodieSparkIOFactory;
 
@@ -438,8 +437,11 @@ public abstract class MultipleSparkJobExecutionStrategy<T>
         .toArray(StoragePath[]::new);
 
     HashMap<String, String> params = new HashMap<>();
-    params.put("hoodie.datasource.query.type", "snapshot");
-    params.put(TIMESTAMP_AS_OF.key(), instantTime);
+    if (hasLogFiles) {
+      params.put("hoodie.datasource.query.type", "snapshot");
+    } else {
+      params.put("hoodie.datasource.query.type", "read_optimized");
+    }
 
     StoragePath[] paths;
     if (hasLogFiles) {

Reply via email to