This is an automated email from the ASF dual-hosted git repository.

leesf pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/master by this push:
     new 142adf4ccb [HUDI-4296] Fix the bug that 
TestHoodieSparkSqlWriter.testSchemaEvolutionForTableType is flaky (#5973)
142adf4ccb is described below

commit 142adf4ccbf92050dd14d1997c0438565fdfc2cd
Author: xiarixiaoyao <[email protected]>
AuthorDate: Sat Jun 25 21:03:19 2022 +0800

    [HUDI-4296] Fix the bug that 
TestHoodieSparkSqlWriter.testSchemaEvolutionForTableType is flaky (#5973)
---
 .../src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala      | 7 ++++---
 1 file changed, 4 insertions(+), 3 deletions(-)

diff --git 
a/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala
 
b/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala
index 4160c34b0c..d6ec645920 100644
--- 
a/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala
+++ 
b/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala
@@ -166,7 +166,9 @@ class BaseFileOnlyRelation(sqlContext: SQLContext,
       DataSource.apply(
         sparkSession = sparkSession,
         paths = extraReadPaths,
-        userSpecifiedSchema = userSchema,
+        // Here we should specify the schema to the latest commit schema since
+        // the table schema evolution.
+        userSpecifiedSchema = userSchema.orElse(Some(tableStructSchema)),
         className = formatClassName,
         // Since we're reading the table as just collection of files we have 
to make sure
         // we only read the latest version of every Hudi's file-group, which 
might be compacted, clustered, etc.
@@ -175,8 +177,7 @@ class BaseFileOnlyRelation(sqlContext: SQLContext,
         // We rely on [[HoodieROTablePathFilter]], to do proper filtering to 
assure that
         options = optParams ++ Map(
           "mapreduce.input.pathFilter.class" -> 
classOf[HoodieROTablePathFilter].getName
-        ),
-        partitionColumns = partitionColumns
+        )
       )
         .resolveRelation()
         .asInstanceOf[HadoopFsRelation]

Reply via email to