This is an automated email from the ASF dual-hosted git repository. yihua pushed a commit to branch branch-0.x in repository https://gitbox.apache.org/repos/asf/hudi.git
commit 3d5d274847ce3782e2d6a9cb94ed8945401c5b16 Author: Y Ethan Guo <[email protected]> AuthorDate: Tue May 14 13:25:42 2024 -0700 [HUDI-7473] Rebalance CI (#10805) --- azure-pipelines-20230430.yml | 173 +++++++++++++++------ .../TestGetPartitionValuesFromPath.scala | 2 +- .../hudi/functional/TestSparkSqlCoreFlow.scala | 7 +- .../apache/hudi/functional/TestSqlStatement.scala | 2 +- .../benchmark/SpaceCurveOptimizeBenchmark.scala | 2 +- .../sql/hudi/command/index/TestIndexSyntax.scala | 2 +- .../hudi/command/index/TestSecondaryIndex.scala | 2 +- .../hudi/{ => common}/HoodieSparkSqlTestBase.scala | 6 +- .../{ => common}/TestHoodieInternalRowUtils.scala | 2 +- .../hudi/{ => common}/TestHoodieOptionConfig.scala | 3 +- .../TestLazyPartitionPathFetching.scala | 2 +- .../TestNestedSchemaPruningOptimization.scala | 2 +- .../TestPartitionPushDownWhenListingPaths.scala | 2 +- .../spark/sql/hudi/{ => common}/TestSqlConf.scala | 5 +- .../spark/sql/hudi/{ => ddl}/TestAlterTable.scala | 4 +- .../{ => ddl}/TestAlterTableDropPartition.scala | 6 +- .../spark/sql/hudi/{ => ddl}/TestCreateTable.scala | 6 +- .../spark/sql/hudi/{ => ddl}/TestSpark3DDL.scala | 4 +- .../sql/hudi/{ => dml}/TestCDCForSparkSQL.scala | 3 +- .../sql/hudi/{ => dml}/TestCompactionTable.scala | 6 +- .../sql/hudi/{ => dml}/TestDataSkippingQuery.scala | 4 +- .../sql/hudi/{ => dml}/TestDeleteFromTable.scala | 4 +- .../spark/sql/hudi/{ => dml}/TestDeleteTable.scala | 3 +- .../spark/sql/hudi/{ => dml}/TestDropTable.scala | 7 +- .../{ => dml}/TestHoodieTableValuedFunction.scala | 3 +- .../spark/sql/hudi/{ => dml}/TestInsertTable.scala | 6 +- .../hudi/{ => dml}/TestMergeIntoLogOnlyTable.scala | 3 +- .../sql/hudi/{ => dml}/TestMergeIntoTable.scala | 6 +- .../sql/hudi/{ => dml}/TestMergeIntoTable2.scala | 3 +- .../TestMergeIntoTableWithNonRecordKeyField.scala | 3 +- .../{ => dml}/TestPartialUpdateForMergeInto.scala | 4 +- .../TestQueryMergeOnReadOptimizedTable.scala | 4 +- .../spark/sql/hudi/{ => dml}/TestRepairTable.scala | 4 +- .../sql/hudi/{ => dml}/TestShowPartitions.scala | 3 +- .../sql/hudi/{ => dml}/TestTimeTravelTable.scala | 3 +- .../sql/hudi/{ => dml}/TestTruncateTable.scala | 3 +- .../spark/sql/hudi/{ => dml}/TestUpdateTable.scala | 3 +- .../procedure/HoodieSparkProcedureTestBase.scala | 2 +- .../sql/hudi/procedure/TestCallCommandParser.scala | 2 +- .../procedure/TestCopyToTempViewProcedure.scala | 2 +- 40 files changed, 207 insertions(+), 106 deletions(-) diff --git a/azure-pipelines-20230430.yml b/azure-pipelines-20230430.yml index b1e3ee5d4d6..e61057a4649 100644 --- a/azure-pipelines-20230430.yml +++ b/azure-pipelines-20230430.yml @@ -30,6 +30,10 @@ parameters: type: object default: - 'hudi-common' + - 'hudi-client/hudi-spark-client' + - name: job2UTModules + type: object + default: - 'hudi-flink-datasource' - 'hudi-flink-datasource/hudi-flink' - 'hudi-flink-datasource/hudi-flink1.14.x' @@ -37,21 +41,20 @@ parameters: - 'hudi-flink-datasource/hudi-flink1.16.x' - 'hudi-flink-datasource/hudi-flink1.17.x' - 'hudi-flink-datasource/hudi-flink1.18.x' - - name: job2Modules + - name: job2FTModules type: object default: + - 'hudi-common' + - 'hudi-flink-datasource' + - 'hudi-flink-datasource/hudi-flink' + - 'hudi-flink-datasource/hudi-flink1.14.x' + - 'hudi-flink-datasource/hudi-flink1.15.x' + - 'hudi-flink-datasource/hudi-flink1.16.x' + - 'hudi-flink-datasource/hudi-flink1.17.x' + - 'hudi-flink-datasource/hudi-flink1.18.x' - 'hudi-client/hudi-spark-client' - 'hudi-spark-datasource/hudi-spark' - - name: job3UTModules - type: object - default: - - 'hudi-spark-datasource' - - 'hudi-spark-datasource/hudi-spark' - - 'hudi-spark-datasource/hudi-spark3.2.x' - - 'hudi-spark-datasource/hudi-spark3.2plus-common' - - 'hudi-spark-datasource/hudi-spark3-common' - - 'hudi-spark-datasource/hudi-spark-common' - - name: job4UTModules + - name: job34UTModules type: object default: - 'hudi-spark-datasource' @@ -60,12 +63,13 @@ parameters: - 'hudi-spark-datasource/hudi-spark3.2plus-common' - 'hudi-spark-datasource/hudi-spark3-common' - 'hudi-spark-datasource/hudi-spark-common' - - name: job5UTModules + - name: job6UTModules type: object default: - '!hudi-hadoop-mr' - '!hudi-client/hudi-java-client' - '!hudi-client/hudi-spark-client' + - '!hudi-cli' - '!hudi-common' - '!hudi-examples' - '!hudi-examples/hudi-examples-common' @@ -85,10 +89,11 @@ parameters: - '!hudi-spark-datasource/hudi-spark3.2plus-common' - '!hudi-spark-datasource/hudi-spark3-common' - '!hudi-spark-datasource/hudi-spark-common' - - name: job5FTModules + - name: job6FTModules type: object default: - '!hudi-client/hudi-spark-client' + - '!hudi-cli' - '!hudi-common' - '!hudi-examples' - '!hudi-examples/hudi-examples-common' @@ -103,18 +108,34 @@ parameters: - '!hudi-flink-datasource/hudi-flink1.17.x' - '!hudi-flink-datasource/hudi-flink1.18.x' - '!hudi-spark-datasource/hudi-spark' + - name: job4HudiSparkDmlOthersWildcardSuites + type: object + default: + - 'org.apache.hudi' + - 'org.apache.spark.hudi' + - 'org.apache.spark.sql.avro' + - 'org.apache.spark.sql.execution' + - 'org.apache.spark.sql.hudi.analysis' + - 'org.apache.spark.sql.hudi.command' + - 'org.apache.spark.sql.hudi.common' + - 'org.apache.spark.sql.hudi.dml' variables: BUILD_PROFILES: '-Dscala-2.12 -Dspark3.2 -Dflink1.18' PLUGIN_OPTS: '-Dcheckstyle.skip=true -Drat.skip=true -Djacoco.skip=true -ntp -B -V -Pwarn-log -Dorg.slf4j.simpleLogger.log.org.apache.maven.plugins.shade=warn -Dorg.slf4j.simpleLogger.log.org.apache.maven.plugins.dependency=warn' MVN_OPTS_INSTALL: '-T 3 -Phudi-platform-service -DskipTests $(BUILD_PROFILES) $(PLUGIN_OPTS) -Dmaven.wagon.httpconnectionManager.ttlSeconds=25 -Dmaven.wagon.http.retryHandler.count=5' MVN_OPTS_TEST: '-fae -Pwarn-log $(BUILD_PROFILES) $(PLUGIN_OPTS)' + JAVA_MVN_TEST_FILTER: '-DwildcardSuites=skipScalaTests -DfailIfNoTests=false' + SCALA_MVN_TEST_FILTER: '-Dtest=skipJavaTests -DfailIfNoTests=false' JOB1_MODULES: ${{ join(',',parameters.job1Modules) }} - JOB2_MODULES: ${{ join(',',parameters.job2Modules) }} - JOB3_MODULES: ${{ join(',',parameters.job3UTModules) }} - JOB4_MODULES: ${{ join(',',parameters.job4UTModules) }} - JOB5_UT_MODULES: ${{ join(',',parameters.job5UTModules) }} - JOB5_FT_MODULES: ${{ join(',',parameters.job5FTModules) }} + JOB2_UT_MODULES: ${{ join(',',parameters.job2UTModules) }} + JOB2_FT_MODULES: ${{ join(',',parameters.job2FTModules) }} + JOB34_MODULES: ${{ join(',',parameters.job34UTModules) }} + JOB3_SPARK_DDL_WILDCARD_SUITES: 'org.apache.spark.sql.hudi.ddl' + JOB6_SPARK_PROCEDURE_WILDCARD_SUITES: 'org.apache.spark.sql.hudi.procedure' + JOB4_SPARK_DML_OTHERS_WILDCARD_SUITES: ${{ join(',',parameters.job4HudiSparkDmlOthersWildcardSuites) }} + JOB6_UT_MODULES: ${{ join(',',parameters.job6UTModules) }} + JOB6_FT_MODULES: ${{ join(',',parameters.job6FTModules) }} stages: - stage: test @@ -123,32 +144,23 @@ stages: value: 1 jobs: - job: UT_FT_1 - displayName: UT FT common & flink & UT client/spark-client - timeoutInMinutes: '150' + displayName: UT common & client/spark-client + timeoutInMinutes: '90' steps: - task: Maven@4 displayName: maven install inputs: mavenPomFile: 'pom.xml' goals: 'clean install' - options: $(MVN_OPTS_INSTALL) - publishJUnitResults: false - jdkVersionOption: '1.8' - - task: Maven@4 - displayName: UT common flink client/spark-client - inputs: - mavenPomFile: 'pom.xml' - goals: 'test' - options: $(MVN_OPTS_TEST) -Punit-tests -pl $(JOB1_MODULES),hudi-client/hudi-spark-client + options: $(MVN_OPTS_INSTALL) -pl $(JOB1_MODULES) -am publishJUnitResults: false jdkVersionOption: '1.8' - mavenOptions: '-Xmx4g' - task: Maven@4 - displayName: FT common flink + displayName: UT common & client/spark-client inputs: mavenPomFile: 'pom.xml' goals: 'test' - options: $(MVN_OPTS_TEST) -Pfunctional-tests -pl $(JOB1_MODULES) + options: $(MVN_OPTS_TEST) -Punit-tests -pl $(JOB1_MODULES) publishJUnitResults: true testResultsFiles: '**/surefire-reports/TEST-*.xml' jdkVersionOption: '1.8' @@ -157,23 +169,32 @@ stages: grep "testcase" */target/surefire-reports/*.xml */*/target/surefire-reports/*.xml | awk -F'"' ' { print $6,$4,$2 } ' | sort -nr | head -n 100 displayName: Top 100 long-running testcases - job: UT_FT_2 - displayName: FT client/spark-client & hudi-spark-datasource/hudi-spark - timeoutInMinutes: '150' + displayName: UT flink & FT common & flink & spark-client & hudi-spark + timeoutInMinutes: '90' steps: - task: Maven@4 displayName: maven install inputs: mavenPomFile: 'pom.xml' goals: 'clean install' - options: $(MVN_OPTS_INSTALL) -pl $(JOB2_MODULES) -am + options: $(MVN_OPTS_INSTALL) -pl $(JOB2_FT_MODULES) -am + publishJUnitResults: false + jdkVersionOption: '1.8' + - task: Maven@4 + displayName: UT flink + inputs: + mavenPomFile: 'pom.xml' + goals: 'test' + options: $(MVN_OPTS_TEST) -Punit-tests -pl $(JOB2_UT_MODULES) publishJUnitResults: false jdkVersionOption: '1.8' + mavenOptions: '-Xmx4g' - task: Maven@4 - displayName: FT client/spark-client & hudi-spark-datasource/hudi-spark + displayName: FT common & flink & client/spark-client & hudi-spark-datasource/hudi-spark inputs: mavenPomFile: 'pom.xml' goals: 'test' - options: $(MVN_OPTS_TEST) -Pfunctional-tests -pl $(JOB2_MODULES) + options: $(MVN_OPTS_TEST) -Pfunctional-tests -pl $(JOB2_FT_MODULES) publishJUnitResults: true testResultsFiles: '**/surefire-reports/TEST-*.xml' jdkVersionOption: '1.8' @@ -182,15 +203,15 @@ stages: grep "testcase" */target/surefire-reports/*.xml */*/target/surefire-reports/*.xml | awk -F'"' ' { print $6,$4,$2 } ' | sort -nr | head -n 100 displayName: Top 100 long-running testcases - job: UT_FT_3 - displayName: Java UT spark-datasource - timeoutInMinutes: '240' + displayName: UT spark-datasource Java Tests & DDL + timeoutInMinutes: '90' steps: - task: Maven@4 displayName: maven install inputs: mavenPomFile: 'pom.xml' goals: 'clean install' - options: $(MVN_OPTS_INSTALL) -pl $(JOB3_MODULES) -am + options: $(MVN_OPTS_INSTALL) -pl $(JOB34_MODULES) -am publishJUnitResults: false jdkVersionOption: '1.8' - task: Maven@4 @@ -198,7 +219,16 @@ stages: inputs: mavenPomFile: 'pom.xml' goals: 'test' - options: $(MVN_OPTS_TEST) -DwildcardSuites=skipScalaTests -DfailIfNoTests=false -Punit-tests -pl $(JOB3_MODULES) + options: $(MVN_OPTS_TEST) -Punit-tests $(JAVA_MVN_TEST_FILTER) -pl $(JOB34_MODULES) + publishJUnitResults: false + jdkVersionOption: '1.8' + mavenOptions: '-Xmx4g' + - task: Maven@4 + displayName: Scala UT spark-datasource DDL + inputs: + mavenPomFile: 'pom.xml' + goals: 'test' + options: $(MVN_OPTS_TEST) -Punit-tests $(SCALA_MVN_TEST_FILTER) -DwildcardSuites="$(JOB3_SPARK_DDL_WILDCARD_SUITES)" -pl $(JOB34_MODULES) publishJUnitResults: true testResultsFiles: '**/surefire-reports/TEST-*.xml' jdkVersionOption: '1.8' @@ -207,23 +237,23 @@ stages: grep "testcase" */target/surefire-reports/*.xml */*/target/surefire-reports/*.xml | awk -F'"' ' { print $6,$4,$2 } ' | sort -nr | head -n 100 displayName: Top 100 long-running testcases - job: UT_FT_4 - displayName: Scala UT spark-datasource - timeoutInMinutes: '240' + displayName: UT spark-datasource DML & others + timeoutInMinutes: '90' steps: - task: Maven@4 displayName: maven install inputs: mavenPomFile: 'pom.xml' goals: 'clean install' - options: $(MVN_OPTS_INSTALL) -pl $(JOB4_MODULES) -am + options: $(MVN_OPTS_INSTALL) -pl $(JOB34_MODULES) -am publishJUnitResults: false jdkVersionOption: '1.8' - task: Maven@4 - displayName: Scala UT spark-datasource + displayName: Scala UT spark-datasource DML & others inputs: mavenPomFile: 'pom.xml' goals: 'test' - options: $(MVN_OPTS_TEST) -Dtest=skipJavaTests -DfailIfNoTests=false -Punit-tests -pl $(JOB4_MODULES) + options: $(MVN_OPTS_TEST) -Punit-tests $(SCALA_MVN_TEST_FILTER) -DwildcardSuites="$(JOB4_SPARK_DML_OTHERS_WILDCARD_SUITES)" -pl $(JOB34_MODULES) publishJUnitResults: true testResultsFiles: '**/surefire-reports/TEST-*.xml' jdkVersionOption: '1.8' @@ -232,8 +262,52 @@ stages: grep "testcase" */target/surefire-reports/*.xml */*/target/surefire-reports/*.xml | awk -F'"' ' { print $6,$4,$2 } ' | sort -nr | head -n 100 displayName: Top 100 long-running testcases - job: UT_FT_5 + displayName: UT FT Hudi Streamer + timeoutInMinutes: '90' + steps: + - task: Docker@2 + displayName: "login to docker hub" + inputs: + command: "login" + containerRegistry: "apachehudi-docker-hub" + - task: Docker@2 + displayName: "load repo into image" + inputs: + containerRegistry: 'apachehudi-docker-hub' + repository: 'apachehudi/hudi-ci-bundle-validation-base' + command: 'build' + Dockerfile: '**/Dockerfile' + ImageName: $(Build.BuildId) + - task: Docker@2 + displayName: "UT FT other modules" + inputs: + containerRegistry: 'apachehudi-docker-hub' + repository: 'apachehudi/hudi-ci-bundle-validation-base' + command: 'run' + arguments: > + -v $(Build.SourcesDirectory):/hudi + -i docker.io/apachehudi/hudi-ci-bundle-validation-base:$(Build.BuildId) + /bin/bash -c "pwd + && rm -rf /hudi/scripts/ci/results + && mvn clean install $(MVN_OPTS_INSTALL) -Phudi-platform-service -Pthrift-gen-source -pl hudi-utilities -am + && mvn test $(MVN_OPTS_TEST) -Punit-tests -Dtest="Test*DeltaStreamer*" -DfailIfNoTests=false -pl hudi-utilities + && mvn test $(MVN_OPTS_TEST) -Pfunctional-tests -Dtest="Test*DeltaStreamer*" -DfailIfNoTests=false -pl hudi-utilities + && ./scripts/ci/move_surefire_reports.sh /hudi /hudi/scripts/ci/results + && echo 'All surefire report files:' + && find . -type f -name \"TEST-*.xml\"" + - task: PublishTestResults@2 + displayName: 'Publish Test Results' + inputs: + testResultsFormat: 'JUnit' + testResultsFiles: '**/surefire-reports/TEST-*.xml' + searchFolder: '$(Build.SourcesDirectory)/scripts/ci/results' + failTaskOnFailedTests: true + - script: | + grep "testcase" scripts/ci/results/*/target/surefire-reports/*.xml scripts/ci/results/*/*/target/surefire-reports/*.xml | awk -F'"' ' { print $6,$4,$2 } ' | sort -nr | head -n 100 + displayName: Top 100 long-running testcases + - job: UT_FT_6 displayName: UT FT other modules - timeoutInMinutes: '240' + timeoutInMinutes: '90' steps: - task: Docker@2 displayName: "login to docker hub" @@ -260,8 +334,9 @@ stages: /bin/bash -c "pwd && rm -rf /hudi/scripts/ci/results && mvn clean install $(MVN_OPTS_INSTALL) -Phudi-platform-service -Pthrift-gen-source - && mvn test $(MVN_OPTS_TEST) -Punit-tests -pl $(JOB5_UT_MODULES) - && mvn test $(MVN_OPTS_TEST) -Pfunctional-tests -pl $(JOB5_FT_MODULES) + && mvn test $(MVN_OPTS_TEST) -Punit-tests $(SCALA_MVN_TEST_FILTER) -DwildcardSuites="$(JOB6_SPARK_PROCEDURE_WILDCARD_SUITES)" -pl $(JOB34_MODULES) + && mvn test $(MVN_OPTS_TEST) -Punit-tests -Dtest="!Test*DeltaStreamer*" -DfailIfNoTests=false -pl $(JOB6_UT_MODULES) + && mvn test $(MVN_OPTS_TEST) -Pfunctional-tests -Dtest="!Test*DeltaStreamer*" -DfailIfNoTests=false -pl $(JOB6_FT_MODULES) && ./scripts/ci/move_surefire_reports.sh /hudi /hudi/scripts/ci/results && echo 'All surefire report files:' && find . -type f -name \"TEST-*.xml\"" diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/TestGetPartitionValuesFromPath.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/TestGetPartitionValuesFromPath.scala index aadd9397f47..9b6feacca0f 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/TestGetPartitionValuesFromPath.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/TestGetPartitionValuesFromPath.scala @@ -18,7 +18,7 @@ package org.apache.hudi.functional -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestGetPartitionValuesFromPath extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/TestSparkSqlCoreFlow.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/TestSparkSqlCoreFlow.scala index b554aa735ec..80d151d5b5e 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/TestSparkSqlCoreFlow.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/TestSparkSqlCoreFlow.scala @@ -28,19 +28,16 @@ import org.apache.hudi.common.table.HoodieTableMetaClient import org.apache.hudi.common.table.timeline.TimelineUtils import org.apache.hudi.common.testutils.HoodieTestDataGenerator import org.apache.hudi.common.testutils.RawTripTestPayload.recordsToStrings +import org.apache.hudi.hadoop.fs.HadoopFSUtils import org.apache.hudi.keygen.NonpartitionedKeyGenerator import org.apache.hudi.{DataSourceReadOptions, HoodieSparkUtils} -import org.apache.hudi.common.fs.FSUtils -import org.apache.hudi.hadoop.fs.HadoopFSUtils - import org.apache.spark.sql -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase import org.apache.spark.sql.{Dataset, Row} import org.junit.jupiter.api.Assertions.{assertEquals, assertTrue} import org.scalatest.Inspectors.forAll import java.io.File - import scala.collection.JavaConversions._ @SparkSQLCoreFlow diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/TestSqlStatement.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/TestSqlStatement.scala index e120cc00fc5..607b99e87b8 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/TestSqlStatement.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/TestSqlStatement.scala @@ -18,7 +18,7 @@ package org.apache.hudi.functional import org.apache.hudi.common.util.FileIOUtils -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestSqlStatement extends HoodieSparkSqlTestBase { val STATE_INIT = 0 diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/execution/benchmark/SpaceCurveOptimizeBenchmark.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/execution/benchmark/SpaceCurveOptimizeBenchmark.scala index 273303fdae6..b185a44dc6f 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/execution/benchmark/SpaceCurveOptimizeBenchmark.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/execution/benchmark/SpaceCurveOptimizeBenchmark.scala @@ -23,7 +23,7 @@ import org.apache.hudi.ColumnStatsIndexHelper.buildColumnStatsTableFor import org.apache.hudi.config.HoodieClusteringConfig.LayoutOptimizationStrategy import org.apache.hudi.sort.SpaceCurveSortingHelper import org.apache.spark.sql.DataFrame -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase import org.apache.spark.sql.types.{IntegerType, StructField} import org.junit.jupiter.api.{Disabled, Tag, Test} diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/command/index/TestIndexSyntax.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/command/index/TestIndexSyntax.scala index cb04c9d8d8b..1b5a52e5ac2 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/command/index/TestIndexSyntax.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/command/index/TestIndexSyntax.scala @@ -22,8 +22,8 @@ package org.apache.spark.sql.hudi.command.index import org.apache.spark.sql.catalyst.analysis.Analyzer import org.apache.spark.sql.catalyst.catalog.CatalogTable import org.apache.spark.sql.catalyst.parser.ParserInterface -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase import org.apache.spark.sql.hudi.command.{CreateIndexCommand, DropIndexCommand, ShowIndexesCommand} +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestIndexSyntax extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/command/index/TestSecondaryIndex.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/command/index/TestSecondaryIndex.scala index eae89099a62..7131cc69e28 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/command/index/TestSecondaryIndex.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/command/index/TestSecondaryIndex.scala @@ -19,7 +19,7 @@ package org.apache.spark.sql.hudi.command.index -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestSecondaryIndex extends HoodieSparkSqlTestBase { test("Test Create/Show/Drop Secondary Index") { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/HoodieSparkSqlTestBase.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/HoodieSparkSqlTestBase.scala similarity index 98% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/HoodieSparkSqlTestBase.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/HoodieSparkSqlTestBase.scala index b9628d05af1..b101e838c84 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/HoodieSparkSqlTestBase.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/HoodieSparkSqlTestBase.scala @@ -15,12 +15,11 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.common import org.apache.hadoop.fs.Path import org.apache.hudi.HoodieSparkRecordMerger import org.apache.hudi.common.config.HoodieStorageConfig -import org.apache.hudi.common.fs.FSUtils import org.apache.hudi.common.model.HoodieAvroRecordMerger import org.apache.hudi.common.model.HoodieRecord.HoodieRecordType import org.apache.hudi.common.table.HoodieTableMetaClient @@ -30,10 +29,9 @@ import org.apache.hudi.exception.ExceptionUtil.getRootCause import org.apache.hudi.hadoop.fs.HadoopFSUtils import org.apache.hudi.index.inmemory.HoodieInMemoryHashIndex import org.apache.hudi.testutils.HoodieClientTestUtils.getSparkConfForTest - import org.apache.spark.SparkConf import org.apache.spark.sql.catalyst.util.DateTimeUtils -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase.checkMessageContains +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase.checkMessageContains import org.apache.spark.sql.{Row, SparkSession} import org.apache.spark.util.Utils import org.joda.time.DateTimeZone diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestHoodieInternalRowUtils.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestHoodieInternalRowUtils.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestHoodieInternalRowUtils.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestHoodieInternalRowUtils.scala index 35afff918b9..2ce4393c6a8 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestHoodieInternalRowUtils.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestHoodieInternalRowUtils.scala @@ -16,7 +16,7 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.common import org.apache.avro.generic.GenericData import org.apache.avro.{LogicalTypes, Schema} diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestHoodieOptionConfig.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestHoodieOptionConfig.scala similarity index 98% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestHoodieOptionConfig.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestHoodieOptionConfig.scala index 43fcb79ecf9..31e5f96d5d8 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestHoodieOptionConfig.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestHoodieOptionConfig.scala @@ -15,11 +15,12 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.common import org.apache.hudi.common.model.{DefaultHoodieRecordPayload, HoodieRecordMerger, OverwriteWithLatestAvroPayload} import org.apache.hudi.common.table.HoodieTableConfig import org.apache.hudi.testutils.SparkClientFunctionalTestHarness +import org.apache.spark.sql.hudi.HoodieOptionConfig import org.apache.spark.sql.types._ import org.junit.jupiter.api.Assertions.assertTrue import org.junit.jupiter.api.Test diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestLazyPartitionPathFetching.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestLazyPartitionPathFetching.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestLazyPartitionPathFetching.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestLazyPartitionPathFetching.scala index e2635c0cba8..aa6cd64fcb3 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestLazyPartitionPathFetching.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestLazyPartitionPathFetching.scala @@ -15,7 +15,7 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.common class TestLazyPartitionPathFetching extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestNestedSchemaPruningOptimization.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestNestedSchemaPruningOptimization.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestNestedSchemaPruningOptimization.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestNestedSchemaPruningOptimization.scala index f8fe24b2174..698d484e16d 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestNestedSchemaPruningOptimization.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestNestedSchemaPruningOptimization.scala @@ -15,7 +15,7 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.common import org.apache.hudi.common.config.HoodieCommonConfig import org.apache.hudi.config.HoodieWriteConfig diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestPartitionPushDownWhenListingPaths.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestPartitionPushDownWhenListingPaths.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestPartitionPushDownWhenListingPaths.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestPartitionPushDownWhenListingPaths.scala index 1b5e590913f..7740da5e664 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestPartitionPushDownWhenListingPaths.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestPartitionPushDownWhenListingPaths.scala @@ -15,7 +15,7 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.common import org.apache.hudi.common.config.HoodieMetadataConfig diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestSqlConf.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestSqlConf.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestSqlConf.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestSqlConf.scala index dbf6d173865..26b21e95437 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestSqlConf.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/common/TestSqlConf.scala @@ -15,7 +15,7 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.common import org.apache.hadoop.conf.Configuration import org.apache.hadoop.fs.Path @@ -23,12 +23,11 @@ import org.apache.hudi.DataSourceReadOptions._ import org.apache.hudi.common.config.DFSPropertiesConfiguration import org.apache.hudi.common.model.HoodieTableType import org.apache.hudi.common.table.{HoodieTableConfig, HoodieTableMetaClient} +import org.scalatest.BeforeAndAfter import java.io.File import java.nio.file.{Files, Paths} -import org.scalatest.BeforeAndAfter - class TestSqlConf extends HoodieSparkSqlTestBase with BeforeAndAfter { def setEnv(key: String, value: String): String = { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestAlterTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/ddl/TestAlterTable.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestAlterTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/ddl/TestAlterTable.scala index b3cd9e497f5..268f5a87bc1 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestAlterTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/ddl/TestAlterTable.scala @@ -15,12 +15,14 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.ddl import org.apache.hudi.HoodieSparkUtils import org.apache.hudi.common.model.HoodieRecord import org.apache.hudi.common.table.{HoodieTableMetaClient, TableSchemaResolver} import org.apache.spark.sql.catalyst.TableIdentifier +import org.apache.spark.sql.hudi.HoodieSqlCommonUtils +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase import org.junit.jupiter.api.Assertions.assertFalse import scala.collection.JavaConverters._ diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestAlterTableDropPartition.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/ddl/TestAlterTableDropPartition.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestAlterTableDropPartition.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/ddl/TestAlterTableDropPartition.scala index 7a146591f4e..f2126da5872 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestAlterTableDropPartition.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/ddl/TestAlterTableDropPartition.scala @@ -15,7 +15,7 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.ddl import org.apache.hudi.DataSourceWriteOptions._ import org.apache.hudi.avro.model.{HoodieCleanMetadata, HoodieCleanPartitionMetadata} @@ -26,8 +26,10 @@ import org.apache.hudi.common.util.{PartitionPathEncodeUtils, StringUtils, Optio import org.apache.hudi.config.{HoodieCleanConfig, HoodieWriteConfig} import org.apache.hudi.keygen.{ComplexKeyGenerator, SimpleKeyGenerator} import org.apache.hudi.{HoodieCLIUtils, HoodieSparkUtils} + import org.apache.spark.sql.SaveMode -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase.getLastCleanMetadata +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase.getLastCleanMetadata import org.junit.jupiter.api.Assertions import org.junit.jupiter.api.Assertions.assertTrue diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestCreateTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/ddl/TestCreateTable.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestCreateTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/ddl/TestCreateTable.scala index 52290ae48b1..0d757f4bedb 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestCreateTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/ddl/TestCreateTable.scala @@ -15,7 +15,7 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.ddl import org.apache.hudi.DataSourceWriteOptions._ import org.apache.hudi.HoodieSparkUtils @@ -28,7 +28,9 @@ import org.apache.hudi.keygen.SimpleKeyGenerator import org.apache.spark.sql.SaveMode import org.apache.spark.sql.catalyst.TableIdentifier import org.apache.spark.sql.catalyst.catalog.{CatalogTableType, HoodieCatalogTable} -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase.getLastCommitMetadata +import org.apache.spark.sql.hudi.HoodieSqlCommonUtils +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase.getLastCommitMetadata import org.apache.spark.sql.types._ import org.junit.jupiter.api.Assertions.{assertFalse, assertTrue} diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestSpark3DDL.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/ddl/TestSpark3DDL.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestSpark3DDL.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/ddl/TestSpark3DDL.scala index bfd14ae4c5a..8ac8e766e56 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestSpark3DDL.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/ddl/TestSpark3DDL.scala @@ -15,7 +15,7 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.ddl import org.apache.hadoop.fs.Path import org.apache.hudi.DataSourceWriteOptions.{PARTITIONPATH_FIELD_OPT_KEY, PRECOMBINE_FIELD_OPT_KEY, RECORDKEY_FIELD_OPT_KEY, SPARK_SQL_INSERT_INTO_OPERATION, TABLE_NAME} @@ -30,6 +30,8 @@ import org.apache.hudi.testutils.DataSourceTestUtils import org.apache.hudi.{DataSourceWriteOptions, HoodieSparkRecordMerger, HoodieSparkUtils} import org.apache.spark.sql.catalyst.TableIdentifier import org.apache.spark.sql.functions.{arrays_zip, col, expr, lit} +import org.apache.spark.sql.hudi.HoodieSqlCommonUtils +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase import org.apache.spark.sql.types.StringType import org.apache.spark.sql.{Row, SaveMode, SparkSession} diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestCDCForSparkSQL.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestCDCForSparkSQL.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestCDCForSparkSQL.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestCDCForSparkSQL.scala index a799ce8f787..59f9eed83b0 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestCDCForSparkSQL.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestCDCForSparkSQL.scala @@ -15,7 +15,7 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.DataSourceReadOptions._ import org.apache.hudi.DataSourceWriteOptions.SPARK_SQL_INSERT_INTO_OPERATION @@ -23,6 +23,7 @@ import org.apache.hudi.common.table.HoodieTableMetaClient import org.apache.hudi.common.table.cdc.HoodieCDCSupplementalLoggingMode.{DATA_BEFORE, DATA_BEFORE_AFTER, OP_KEY_ONLY} import org.apache.spark.sql.DataFrame import org.apache.spark.sql.functions._ +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase import org.junit.jupiter.api.Assertions.assertEquals class TestCDCForSparkSQL extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestCompactionTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestCompactionTable.scala similarity index 97% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestCompactionTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestCompactionTable.scala index 5ded75dcdab..31948c3298d 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestCompactionTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestCompactionTable.scala @@ -15,12 +15,14 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml + +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestCompactionTable extends HoodieSparkSqlTestBase { test("Test compaction table") { - withRecordType()(withTempDir {tmp => + withRecordType()(withTempDir { tmp => val tableName = generateTableName spark.sql( s""" diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestDataSkippingQuery.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestDataSkippingQuery.scala similarity index 98% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestDataSkippingQuery.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestDataSkippingQuery.scala index 1ac7185f642..23255b763ff 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestDataSkippingQuery.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestDataSkippingQuery.scala @@ -17,7 +17,9 @@ * under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml + +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestDataSkippingQuery extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestDeleteFromTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestDeleteFromTable.scala similarity index 96% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestDeleteFromTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestDeleteFromTable.scala index e3ea0173022..b289ce74646 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestDeleteFromTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestDeleteFromTable.scala @@ -15,7 +15,9 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml + +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestDeleteFromTable extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestDeleteTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestDeleteTable.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestDeleteTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestDeleteTable.scala index bc87405b9f9..b9cafb6ec07 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestDeleteTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestDeleteTable.scala @@ -15,12 +15,13 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.DataSourceWriteOptions._ import org.apache.hudi.HoodieSparkUtils.isSpark2 import org.apache.hudi.config.HoodieWriteConfig import org.apache.spark.sql.SaveMode +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestDeleteTable extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestDropTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestDropTable.scala similarity index 98% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestDropTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestDropTable.scala index 0781fc6af06..743abc5b2fd 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestDropTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestDropTable.scala @@ -15,15 +15,14 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml -import org.apache.hudi.common.fs.FSUtils +import org.apache.hadoop.fs.Path import org.apache.hudi.hadoop.fs.HadoopFSUtils - -import org.apache.hadoop.fs.{LocalFileSystem, Path} import org.apache.spark.sql.AnalysisException import org.apache.spark.sql.catalyst.TableIdentifier import org.apache.spark.sql.catalyst.catalog.SessionCatalog +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestDropTable extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestHoodieTableValuedFunction.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestHoodieTableValuedFunction.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestHoodieTableValuedFunction.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestHoodieTableValuedFunction.scala index 1809a7e2f44..58f052df8f3 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestHoodieTableValuedFunction.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestHoodieTableValuedFunction.scala @@ -15,11 +15,12 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.DataSourceWriteOptions.SPARK_SQL_INSERT_INTO_OPERATION import org.apache.hudi.HoodieSparkUtils import org.apache.spark.sql.functions.{col, from_json} +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestHoodieTableValuedFunction extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestInsertTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestInsertTable.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestInsertTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestInsertTable.scala index 38f2e4e428c..b2261447181 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestInsertTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestInsertTable.scala @@ -15,7 +15,7 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.DataSourceWriteOptions._ import org.apache.hudi.common.model.HoodieRecord.HoodieRecordType @@ -29,8 +29,10 @@ import org.apache.hudi.execution.bulkinsert.BulkInsertSortMode import org.apache.hudi.index.HoodieIndex.IndexType import org.apache.hudi.{DataSourceWriteOptions, HoodieCLIUtils, HoodieSparkUtils} import org.apache.spark.sql.SaveMode -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase.getLastCommitMetadata +import org.apache.spark.sql.hudi.HoodieSqlCommonUtils import org.apache.spark.sql.hudi.command.HoodieSparkValidateDuplicateKeyRecordMerger +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase.getLastCommitMetadata import org.junit.jupiter.api.Assertions.assertEquals import java.io.File diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestMergeIntoLogOnlyTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestMergeIntoLogOnlyTable.scala similarity index 97% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestMergeIntoLogOnlyTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestMergeIntoLogOnlyTable.scala index 48ee872d4d9..d25b9752e35 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestMergeIntoLogOnlyTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestMergeIntoLogOnlyTable.scala @@ -15,9 +15,10 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.testutils.DataSourceTestUtils +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestMergeIntoLogOnlyTable extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestMergeIntoTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestMergeIntoTable.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestMergeIntoTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestMergeIntoTable.scala index b56ca09ab96..7fe9a753014 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestMergeIntoTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestMergeIntoTable.scala @@ -15,13 +15,15 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.DataSourceWriteOptions.SPARK_SQL_OPTIMIZED_WRITES +import org.apache.hudi.config.HoodieWriteConfig.MERGE_SMALL_FILE_GROUP_CANDIDATES_LIMIT import org.apache.hudi.{DataSourceReadOptions, HoodieDataSourceHelpers, HoodieSparkUtils, ScalaAssertionSupport} import org.apache.hudi.common.fs.FSUtils import org.apache.hudi.hadoop.fs.HadoopFSUtils - +import org.apache.hudi.{DataSourceReadOptions, HoodieDataSourceHelpers, HoodieSparkUtils, ScalaAssertionSupport} +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase import org.apache.spark.sql.internal.SQLConf class TestMergeIntoTable extends HoodieSparkSqlTestBase with ScalaAssertionSupport { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestMergeIntoTable2.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestMergeIntoTable2.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestMergeIntoTable2.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestMergeIntoTable2.scala index 8ea7284e840..f58935b5bf3 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestMergeIntoTable2.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestMergeIntoTable2.scala @@ -15,11 +15,12 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.HoodieSparkUtils import org.apache.hudi.common.table.HoodieTableMetaClient import org.apache.spark.sql.Row +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestMergeIntoTable2 extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestMergeIntoTableWithNonRecordKeyField.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestMergeIntoTableWithNonRecordKeyField.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestMergeIntoTableWithNonRecordKeyField.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestMergeIntoTableWithNonRecordKeyField.scala index dae2dda4bfa..8e06995475b 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestMergeIntoTableWithNonRecordKeyField.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestMergeIntoTableWithNonRecordKeyField.scala @@ -15,10 +15,11 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.DataSourceWriteOptions.SPARK_SQL_OPTIMIZED_WRITES import org.apache.hudi.{HoodieSparkUtils, ScalaAssertionSupport} +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestMergeIntoTableWithNonRecordKeyField extends HoodieSparkSqlTestBase with ScalaAssertionSupport { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestPartialUpdateForMergeInto.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestPartialUpdateForMergeInto.scala similarity index 97% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestPartialUpdateForMergeInto.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestPartialUpdateForMergeInto.scala index 2284d76ab3a..e83270930f4 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestPartialUpdateForMergeInto.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestPartialUpdateForMergeInto.scala @@ -15,10 +15,12 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.HoodieSparkUtils +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase + class TestPartialUpdateForMergeInto extends HoodieSparkSqlTestBase { test("Test Partial Update") { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestQueryMergeOnReadOptimizedTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestQueryMergeOnReadOptimizedTable.scala similarity index 96% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestQueryMergeOnReadOptimizedTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestQueryMergeOnReadOptimizedTable.scala index 3f6934d9734..f5c9433a60e 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestQueryMergeOnReadOptimizedTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestQueryMergeOnReadOptimizedTable.scala @@ -15,7 +15,9 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml + +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestQueryMergeOnReadOptimizedTable extends HoodieSparkSqlTestBase { test("Test Query Merge_On_Read Read_Optimized table") { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestRepairTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestRepairTable.scala similarity index 98% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestRepairTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestRepairTable.scala index 8078ed29bd7..fccc7b61f1f 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestRepairTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestRepairTable.scala @@ -16,14 +16,14 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.DataSourceWriteOptions.{PARTITIONPATH_FIELD, PRECOMBINE_FIELD, RECORDKEY_FIELD} import org.apache.hudi.HoodieSparkUtils import org.apache.hudi.common.table.HoodieTableConfig.HIVE_STYLE_PARTITIONING_ENABLE import org.apache.hudi.config.HoodieWriteConfig.TBL_NAME - import org.apache.spark.sql.SaveMode +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestRepairTable extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestShowPartitions.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestShowPartitions.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestShowPartitions.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestShowPartitions.scala index 968d7a168aa..ff8168c5191 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestShowPartitions.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestShowPartitions.scala @@ -15,10 +15,11 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.HoodieSparkUtils.isSpark2 import org.apache.hudi.common.util.PartitionPathEncodeUtils.DEFAULT_PARTITION_PATH +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestShowPartitions extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestTimeTravelTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestTimeTravelTable.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestTimeTravelTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestTimeTravelTable.scala index e6275d22e62..9924b700353 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestTimeTravelTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestTimeTravelTable.scala @@ -15,10 +15,11 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.HoodieSparkUtils import org.apache.hudi.common.table.HoodieTableMetaClient +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestTimeTravelTable extends HoodieSparkSqlTestBase { test("Test Insert and Update Record with time travel") { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestTruncateTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestTruncateTable.scala similarity index 98% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestTruncateTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestTruncateTable.scala index 808bfebb802..411562c3558 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestTruncateTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestTruncateTable.scala @@ -16,11 +16,12 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.DataSourceWriteOptions._ import org.apache.hudi.config.HoodieWriteConfig import org.apache.spark.sql.SaveMode +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestTruncateTable extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestUpdateTable.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestUpdateTable.scala similarity index 99% rename from hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestUpdateTable.scala rename to hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestUpdateTable.scala index 7c7fc70d3f3..5d023b8d856 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/TestUpdateTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/dml/TestUpdateTable.scala @@ -15,12 +15,13 @@ * limitations under the License. */ -package org.apache.spark.sql.hudi +package org.apache.spark.sql.hudi.dml import org.apache.hudi.DataSourceWriteOptions.SPARK_SQL_OPTIMIZED_WRITES import org.apache.hudi.HoodieSparkUtils.isSpark2 import org.apache.hudi.common.model.HoodieTableType import org.apache.hudi.common.table.HoodieTableMetaClient +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase import org.junit.jupiter.api.Assertions.assertEquals class TestUpdateTable extends HoodieSparkSqlTestBase { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/HoodieSparkProcedureTestBase.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/HoodieSparkProcedureTestBase.scala index cff41105117..ff4f7aa6ab0 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/HoodieSparkProcedureTestBase.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/HoodieSparkProcedureTestBase.scala @@ -19,7 +19,7 @@ package org.apache.spark.sql.hudi.procedure import org.apache.spark.sql.Dataset import org.apache.spark.sql.execution.columnar.InMemoryRelation -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class HoodieSparkProcedureTestBase extends HoodieSparkSqlTestBase { override def generateTableName: String = { diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/TestCallCommandParser.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/TestCallCommandParser.scala index b5b13f46806..3d07286ca19 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/TestCallCommandParser.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/TestCallCommandParser.scala @@ -21,7 +21,7 @@ import org.apache.hudi.HoodieSparkUtils import org.apache.hudi.common.util.CollectionUtils.createImmutableList import org.apache.spark.sql.catalyst.expressions.Literal import org.apache.spark.sql.catalyst.plans.logical.{CallCommand, NamedArgument, PositionalArgument} -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase import org.apache.spark.sql.types.{DataType, DataTypes} import java.math.BigDecimal diff --git a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/TestCopyToTempViewProcedure.scala b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/TestCopyToTempViewProcedure.scala index 5cb5b68fa04..6f54dfb5094 100644 --- a/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/TestCopyToTempViewProcedure.scala +++ b/hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/spark/sql/hudi/procedure/TestCopyToTempViewProcedure.scala @@ -18,7 +18,7 @@ package org.apache.spark.sql.hudi.procedure import org.apache.hudi.HoodieSparkUtils -import org.apache.spark.sql.hudi.HoodieSparkSqlTestBase +import org.apache.spark.sql.hudi.common.HoodieSparkSqlTestBase class TestCopyToTempViewProcedure extends HoodieSparkSqlTestBase {
