This is an automated email from the ASF dual-hosted git repository.
yihua pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/master by this push:
new 56aded81287 [HUDI-7599] add bootstrap mor legacy reader back to
default source (#10990)
56aded81287 is described below
commit 56aded81287f295cfee692f16be0adc6f175902e
Author: Jon Vexler <[email protected]>
AuthorDate: Fri Apr 12 14:31:10 2024 -0400
[HUDI-7599] add bootstrap mor legacy reader back to default source (#10990)
Co-authored-by: Jonathan Vexler <=>
---
.../src/main/scala/org/apache/hudi/DefaultSource.scala | 12 ++++++++++--
.../hudi/functional/TestNewHoodieParquetFileFormat.java | 12 +++---------
2 files changed, 13 insertions(+), 11 deletions(-)
diff --git
a/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/DefaultSource.scala
b/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/DefaultSource.scala
index be3d2f4ed4b..8efa8e28867 100644
---
a/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/DefaultSource.scala
+++
b/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/DefaultSource.scala
@@ -299,14 +299,22 @@ object DefaultSource {
new IncrementalRelation(sqlContext, parameters, userSchema,
metaClient)
}
- case (MERGE_ON_READ, QUERY_TYPE_SNAPSHOT_OPT_VAL, _) =>
+ case (MERGE_ON_READ, QUERY_TYPE_SNAPSHOT_OPT_VAL, false) =>
if (useNewParquetFileFormat) {
new HoodieMergeOnReadSnapshotHadoopFsRelationFactory(
- sqlContext, metaClient, parameters, userSchema,
isBootstrappedTable).build()
+ sqlContext, metaClient, parameters, userSchema, isBootstrap =
false).build()
} else {
new MergeOnReadSnapshotRelation(sqlContext, parameters,
metaClient, globPaths, userSchema)
}
+ case (MERGE_ON_READ, QUERY_TYPE_SNAPSHOT_OPT_VAL, true) =>
+ if (useNewParquetFileFormat) {
+ new HoodieMergeOnReadSnapshotHadoopFsRelationFactory(
+ sqlContext, metaClient, parameters, userSchema, isBootstrap =
true).build()
+ } else {
+ HoodieBootstrapMORRelation(sqlContext, userSchema, globPaths,
metaClient, parameters)
+ }
+
case (MERGE_ON_READ, QUERY_TYPE_INCREMENTAL_OPT_VAL, _) =>
if (useNewParquetFileFormat) {
new HoodieMergeOnReadIncrementalHadoopFsRelationFactory(
diff --git
a/hudi-spark-datasource/hudi-spark/src/test/java/org/apache/hudi/functional/TestNewHoodieParquetFileFormat.java
b/hudi-spark-datasource/hudi-spark/src/test/java/org/apache/hudi/functional/TestNewHoodieParquetFileFormat.java
index ce462c93d1b..be2b6ff949e 100644
---
a/hudi-spark-datasource/hudi-spark/src/test/java/org/apache/hudi/functional/TestNewHoodieParquetFileFormat.java
+++
b/hudi-spark-datasource/hudi-spark/src/test/java/org/apache/hudi/functional/TestNewHoodieParquetFileFormat.java
@@ -24,7 +24,6 @@ import org.apache.hudi.common.model.HoodieTableType;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.Row;
import org.apache.spark.sql.SaveMode;
-import org.junit.jupiter.api.Disabled;
import org.junit.jupiter.api.Tag;
import org.junit.jupiter.params.ParameterizedTest;
import org.junit.jupiter.params.provider.Arguments;
@@ -39,18 +38,13 @@ import static
org.apache.hudi.common.model.HoodieTableType.MERGE_ON_READ;
import static org.junit.jupiter.api.Assertions.assertEquals;
@Tag("functional")
-@Disabled("HUDI-6756")
public class TestNewHoodieParquetFileFormat extends TestBootstrapReadBase {
private static Stream<Arguments> testArgs() {
Stream.Builder<Arguments> b = Stream.builder();
- HoodieTableType[] tableType = {COPY_ON_WRITE, MERGE_ON_READ};
- Integer[] nPartitions = {0, 1, 2};
- for (HoodieTableType tt : tableType) {
- for (Integer n : nPartitions) {
- b.add(Arguments.of(tt, n));
- }
- }
+ b.add(Arguments.of(MERGE_ON_READ, 0));
+ b.add(Arguments.of(COPY_ON_WRITE, 1));
+ b.add(Arguments.of(MERGE_ON_READ, 2));
return b.build();
}