This is an automated email from the ASF dual-hosted git repository.
danny0405 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/master by this push:
new 98b3d3bac0f [HUDI-7980] Optimize the configuration content when
performing clustering with row writer (#11614)
98b3d3bac0f is described below
commit 98b3d3bac0f31219e5b93b7528516b27b87ea699
Author: majian <[email protected]>
AuthorDate: Sun Jul 14 09:06:37 2024 +0800
[HUDI-7980] Optimize the configuration content when performing clustering
with row writer (#11614)
---
.../run/strategy/MultipleSparkJobExecutionStrategy.java | 8 +++++---
1 file changed, 5 insertions(+), 3 deletions(-)
diff --git
a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/clustering/run/strategy/MultipleSparkJobExecutionStrategy.java
b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/clustering/run/strategy/MultipleSparkJobExecutionStrategy.java
index 75b42491eda..47ccd8700a8 100644
---
a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/clustering/run/strategy/MultipleSparkJobExecutionStrategy.java
+++
b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/clustering/run/strategy/MultipleSparkJobExecutionStrategy.java
@@ -93,7 +93,6 @@ import java.util.stream.Collectors;
import java.util.stream.Stream;
import static
org.apache.hudi.client.utils.SparkPartitionUtils.getPartitionFieldVals;
-import static org.apache.hudi.common.config.HoodieCommonConfig.TIMESTAMP_AS_OF;
import static
org.apache.hudi.config.HoodieClusteringConfig.PLAN_STRATEGY_SORT_COLUMNS;
import static
org.apache.hudi.io.storage.HoodieSparkIOFactory.getHoodieSparkIOFactory;
@@ -438,8 +437,11 @@ public abstract class MultipleSparkJobExecutionStrategy<T>
.toArray(StoragePath[]::new);
HashMap<String, String> params = new HashMap<>();
- params.put("hoodie.datasource.query.type", "snapshot");
- params.put(TIMESTAMP_AS_OF.key(), instantTime);
+ if (hasLogFiles) {
+ params.put("hoodie.datasource.query.type", "snapshot");
+ } else {
+ params.put("hoodie.datasource.query.type", "read_optimized");
+ }
StoragePath[] paths;
if (hasLogFiles) {