This is an automated email from the ASF dual-hosted git repository.
xushiyan pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/master by this push:
new 7592ddd [HUDI-2399] Rebalance CI jobs for shorter wait time (#3604)
7592ddd is described below
commit 7592ddd776148bf750043d31d74211670b961c46
Author: Raymond Xu <[email protected]>
AuthorDate: Sun Sep 5 09:25:57 2021 -0700
[HUDI-2399] Rebalance CI jobs for shorter wait time (#3604)
---
azure-pipelines.yml | 133 +++++++++++++++++++++++++++++++++++++++-------------
1 file changed, 100 insertions(+), 33 deletions(-)
diff --git a/azure-pipelines.yml b/azure-pipelines.yml
index d290a7c..a0bf6a7 100644
--- a/azure-pipelines.yml
+++ b/azure-pipelines.yml
@@ -23,92 +23,158 @@ pool:
variables:
MAVEN_CACHE_FOLDER: $(Pipeline.Workspace)/.m2/repository
- MAVEN_OPTS: '-Dmaven.repo.local=$(MAVEN_CACHE_FOLDER) -Dcheckstyle.skip=true
-Drat.skip=true'
+ MAVEN_OPTS: '-Dmaven.repo.local=$(MAVEN_CACHE_FOLDER) -Dcheckstyle.skip=true
-Drat.skip=true -Djacoco.skip=true'
SPARK_VERSION: '2.4.4'
HADOOP_VERSION: '2.7'
SPARK_HOME:
$(Pipeline.Workspace)/spark-$(SPARK_VERSION)-bin-hadoop$(HADOOP_VERSION)
stages:
- stage: test
- displayName: "Run tests"
jobs:
- - job: unit_tests_spark_client
+ - job: UT_FT_1
+ displayName: UT FT common & flink & UT client/spark-client
steps:
- task: Cache@2
+ displayName: set cache
inputs:
key: 'maven | "$(Agent.OS)" | **/pom.xml'
restoreKeys: |
maven | "$(Agent.OS)"
maven
path: $(MAVEN_CACHE_FOLDER)
- displayName: Cache Maven local repo
- - script: |
- mvn $(MAVEN_OPTS) clean install -DskipTests
- task: Maven@3
+ displayName: maven install
+ inputs:
+ mavenPomFile: 'pom.xml'
+ goals: 'install'
+ options: -DskipTests
+ publishJUnitResults: false
+ jdkVersionOption: '1.8'
+ mavenOptions: '-Xmx2g $(MAVEN_OPTS)'
+ - task: Maven@3
+ displayName: UT common flink client/spark-client
+ inputs:
+ mavenPomFile: 'pom.xml'
+ goals: 'test'
+ options: -Punit-tests -pl
hudi-common,hudi-flink,hudi-client/hudi-spark-client
+ publishJUnitResults: false
+ jdkVersionOption: '1.8'
+ mavenOptions: '-Xmx2g $(MAVEN_OPTS)'
+ - task: Maven@3
+ displayName: FT common flink
inputs:
mavenPomFile: 'pom.xml'
goals: 'test'
- options: -Punit-tests -pl hudi-client/hudi-spark-client
+ options: -Pfunctional-tests -pl hudi-common,hudi-flink
publishJUnitResults: false
- testResultsFiles: '**/surefire-reports/TEST-*.xml'
- testRunTitle: 'unit tests spark client'
- javaHomeOption: 'JDKVersion'
jdkVersionOption: '1.8'
- jdkArchitectureOption: 'x64'
mavenOptions: '-Xmx2g $(MAVEN_OPTS)'
- - job: unit_tests_utilities
+ - job: UT_FT_2
+ displayName: FT client/spark-client
steps:
- task: Cache@2
+ displayName: set cache
inputs:
key: 'maven | "$(Agent.OS)" | **/pom.xml'
restoreKeys: |
maven | "$(Agent.OS)"
maven
path: $(MAVEN_CACHE_FOLDER)
- displayName: Cache Maven local repo
- - script: |
- mvn $(MAVEN_OPTS) clean install -DskipTests
- task: Maven@3
+ displayName: maven install
+ inputs:
+ mavenPomFile: 'pom.xml'
+ goals: 'install'
+ options: -DskipTests
+ publishJUnitResults: false
+ jdkVersionOption: '1.8'
+ mavenOptions: '-Xmx2g $(MAVEN_OPTS)'
+ - task: Maven@3
+ displayName: FT client/spark-client
inputs:
mavenPomFile: 'pom.xml'
goals: 'test'
- options: -Punit-tests -pl hudi-utilities
+ options: -Pfunctional-tests -pl hudi-client/hudi-spark-client
publishJUnitResults: false
- testResultsFiles: '**/surefire-reports/TEST-*.xml'
- testRunTitle: 'unit tests utilities'
- javaHomeOption: 'JDKVersion'
jdkVersionOption: '1.8'
- jdkArchitectureOption: 'x64'
mavenOptions: '-Xmx2g $(MAVEN_OPTS)'
- - job: unit_tests_other_modules
+ - job: UT_FT_3
+ displayName: UT FT cli & utilities & sync/hive-sync
steps:
- task: Cache@2
+ displayName: set cache
inputs:
key: 'maven | "$(Agent.OS)" | **/pom.xml'
restoreKeys: |
maven | "$(Agent.OS)"
maven
path: $(MAVEN_CACHE_FOLDER)
- displayName: Cache Maven local repo
- - script: |
- mvn $(MAVEN_OPTS) clean install -DskipTests
- task: Maven@3
+ displayName: maven install
+ inputs:
+ mavenPomFile: 'pom.xml'
+ goals: 'install'
+ options: -DskipTests
+ publishJUnitResults: false
+ jdkVersionOption: '1.8'
+ mavenOptions: '-Xmx2g $(MAVEN_OPTS)'
+ - task: Maven@3
+ displayName: UT cli & utilities & sync/hive-sync
inputs:
mavenPomFile: 'pom.xml'
goals: 'test'
- options: -Punit-tests -pl
!hudi-utilities,!hudi-client/hudi-spark-client
+ options: -Punit-tests -pl
hudi-cli,hudi-utilities,hudi-sync/hudi-hive-sync
publishJUnitResults: false
- testResultsFiles: '**/surefire-reports/TEST-*.xml'
- testRunTitle: 'unit tests other modules'
- javaHomeOption: 'JDKVersion'
jdkVersionOption: '1.8'
- jdkArchitectureOption: 'x64'
mavenOptions: '-Xmx2g $(MAVEN_OPTS)'
- - job: functional_tests
+ - task: Maven@3
+ displayName: FT cli & utilities & sync/hive-sync
+ inputs:
+ mavenPomFile: 'pom.xml'
+ goals: 'test'
+ options: -Pfunctional-tests -pl
hudi-cli,hudi-utilities,hudi-sync/hudi-hive-sync
+ publishJUnitResults: false
+ jdkVersionOption: '1.8'
+ mavenOptions: '-Xmx2g $(MAVEN_OPTS)'
+ - job: UT_FT_4
+ displayName: UT FT other modules
steps:
- - script: |
- mvn $(MAVEN_OPTS) -Pfunctional-tests test
- - job: integration_tests
+ - task: Cache@2
+ displayName: set cache
+ inputs:
+ key: 'maven | "$(Agent.OS)" | **/pom.xml'
+ restoreKeys: |
+ maven | "$(Agent.OS)"
+ maven
+ path: $(MAVEN_CACHE_FOLDER)
+ - task: Maven@3
+ displayName: maven install
+ inputs:
+ mavenPomFile: 'pom.xml'
+ goals: 'install'
+ options: -DskipTests
+ publishJUnitResults: false
+ jdkVersionOption: '1.8'
+ mavenOptions: '-Xmx2g $(MAVEN_OPTS)'
+ - task: Maven@3
+ displayName: UT other modules
+ inputs:
+ mavenPomFile: 'pom.xml'
+ goals: 'test'
+ options: -Punit-tests -pl
!hudi-common,!hudi-flink,!hudi-client/hudi-spark-client,!hudi-cli,!hudi-utilities,!hudi-sync/hudi-hive-sync
+ publishJUnitResults: false
+ jdkVersionOption: '1.8'
+ mavenOptions: '-Xmx2g $(MAVEN_OPTS)'
+ - task: Maven@3
+ displayName: FT other modules
+ inputs:
+ mavenPomFile: 'pom.xml'
+ goals: 'test'
+ options: -Pfunctional-tests -pl
!hudi-common,!hudi-flink,!hudi-client/hudi-spark-client,!hudi-cli,!hudi-utilities,!hudi-sync/hudi-hive-sync
+ publishJUnitResults: false
+ jdkVersionOption: '1.8'
+ mavenOptions: '-Xmx2g $(MAVEN_OPTS)'
+ - job: IT
steps:
- script: |
echo 'Downloading
spark-$(SPARK_VERSION)-bin-hadoop$(HADOOP_VERSION)'
@@ -116,3 +182,4 @@ stages:
tar -xvf $(Pipeline.Workspace)/spark-$(SPARK_VERSION).tgz -C
$(Pipeline.Workspace)/
mkdir /tmp/spark-events/
mvn $(MAVEN_OPTS) -Pintegration-tests verify
+ displayName: IT