This is an automated email from the ASF dual-hosted git repository. yihua pushed a commit to branch release-0.14.1-spark35-scala213 in repository https://gitbox.apache.org/repos/asf/hudi.git
commit 347093eab82af2b9dc54baae57ee86cc73455b2b Author: Y Ethan Guo <[email protected]> AuthorDate: Thu Feb 22 17:00:41 2024 -0800 [MINOR] Move release candidate validation to a separate GitHub action (#10729) --- .github/workflows/bot.yml | 69 -------------- .github/workflows/release_candidate_validation.yml | 100 +++++++++++++++++++++ 2 files changed, 100 insertions(+), 69 deletions(-) diff --git a/.github/workflows/bot.yml b/.github/workflows/bot.yml index a82ba7d79fe..1bd8f71efc5 100644 --- a/.github/workflows/bot.yml +++ b/.github/workflows/bot.yml @@ -368,75 +368,6 @@ jobs: HUDI_VERSION=$(mvn help:evaluate -Dexpression=project.version -q -DforceStdout) ./packaging/bundle-validation/ci_run.sh $HUDI_VERSION openjdk17 - validate-release-candidate-bundles: - if: false - runs-on: ubuntu-latest - env: - HUDI_VERSION: 0.13.1-rcx - STAGING_REPO_NUM: 1123 - strategy: - matrix: - include: - - flinkProfile: 'flink1.18' - sparkProfile: 'spark3' - sparkRuntime: 'spark3.5.0' - - flinkProfile: 'flink1.18' - sparkProfile: 'spark3.5' - sparkRuntime: 'spark3.5.0' - - flinkProfile: 'flink1.18' - sparkProfile: 'spark3.4' - sparkRuntime: 'spark3.4.0' - - flinkProfile: 'flink1.17' - sparkProfile: 'spark3.3' - sparkRuntime: 'spark3.3.2' - - flinkProfile: 'flink1.16' - sparkProfile: 'spark3.3' - sparkRuntime: 'spark3.3.1' - - flinkProfile: 'flink1.15' - sparkProfile: 'spark3.2' - sparkRuntime: 'spark3.2.3' - - flinkProfile: 'flink1.13' - sparkProfile: 'spark3.1' - sparkRuntime: 'spark3.1.3' - - flinkProfile: 'flink1.13' - sparkProfile: 'spark' - sparkRuntime: 'spark2.4.8' - - flinkProfile: 'flink1.13' - sparkProfile: 'spark2.4' - sparkRuntime: 'spark2.4.8' - steps: - - uses: actions/checkout@v3 - - name: Set up JDK 8 - uses: actions/setup-java@v3 - with: - java-version: '8' - distribution: 'adopt' - architecture: x64 - cache: maven - - name: IT - Bundle Validation - OpenJDK 8 - env: - FLINK_PROFILE: ${{ matrix.flinkProfile }} - SPARK_PROFILE: ${{ matrix.sparkProfile }} - SPARK_RUNTIME: ${{ matrix.sparkRuntime }} - run: | - ./packaging/bundle-validation/ci_run.sh $HUDI_VERSION openjdk8 $STAGING_REPO_NUM - - name: IT - Bundle Validation - OpenJDK 11 - env: - FLINK_PROFILE: ${{ matrix.flinkProfile }} - SPARK_PROFILE: ${{ matrix.sparkProfile }} - SPARK_RUNTIME: ${{ matrix.sparkRuntime }} - if: ${{ startsWith(env.SPARK_PROFILE, 'spark3') }} # Only Spark 3.x supports Java 11 as of now - run: | - ./packaging/bundle-validation/ci_run.sh $HUDI_VERSION openjdk11 $STAGING_REPO_NUM - - name: IT - Bundle Validation - OpenJDK 17 - env: - FLINK_PROFILE: ${{ matrix.flinkProfile }} - SPARK_PROFILE: ${{ matrix.sparkProfile }} - SPARK_RUNTIME: ${{ matrix.sparkRuntime }} - if: ${{ endsWith(env.SPARK_PROFILE, '3.3') }} # Only Spark 3.3 supports Java 17 as of now - run: | - ./packaging/bundle-validation/ci_run.sh $HUDI_VERSION openjdk17 $STAGING_REPO_NUM - integration-tests: runs-on: ubuntu-latest strategy: diff --git a/.github/workflows/release_candidate_validation.yml b/.github/workflows/release_candidate_validation.yml new file mode 100644 index 00000000000..2f14fd96f7d --- /dev/null +++ b/.github/workflows/release_candidate_validation.yml @@ -0,0 +1,100 @@ +name: Release Candidate Validation + +on: + push: + branches: + - 'release-*' + pull_request: + paths-ignore: + - '**.bmp' + - '**.gif' + - '**.jpg' + - '**.jpeg' + - '**.md' + - '**.pdf' + - '**.png' + - '**.svg' + - '**.yaml' + - '.gitignore' + branches: + - 'release-*' + +concurrency: + group: ${{ github.ref }} + cancel-in-progress: ${{ !contains(github.ref, 'master') }} + +env: + MVN_ARGS: -e -ntp -B -V -Dgpg.skip -Djacoco.skip -Pwarn-log -Dorg.slf4j.simpleLogger.log.org.apache.maven.plugins.shade=warn -Dorg.slf4j.simpleLogger.log.org.apache.maven.plugins.dependency=warn -Dmaven.wagon.httpconnectionManager.ttlSeconds=25 -Dmaven.wagon.http.retryHandler.count=5 + SPARK_COMMON_MODULES: hudi-spark-datasource/hudi-spark,hudi-spark-datasource/hudi-spark-common + +jobs: + validate-release-candidate-bundles: + runs-on: ubuntu-latest + env: + HUDI_VERSION: 0.14.1 + STAGING_REPO_NUM: 1123 + strategy: + matrix: + include: + - flinkProfile: 'flink1.18' + sparkProfile: 'spark3' + sparkRuntime: 'spark3.5.0' + - flinkProfile: 'flink1.18' + sparkProfile: 'spark3.5' + sparkRuntime: 'spark3.5.0' + - flinkProfile: 'flink1.18' + sparkProfile: 'spark3.4' + sparkRuntime: 'spark3.4.0' + - flinkProfile: 'flink1.17' + sparkProfile: 'spark3.3' + sparkRuntime: 'spark3.3.2' + - flinkProfile: 'flink1.16' + sparkProfile: 'spark3.3' + sparkRuntime: 'spark3.3.1' + - flinkProfile: 'flink1.15' + sparkProfile: 'spark3.2' + sparkRuntime: 'spark3.2.3' + - flinkProfile: 'flink1.14' + sparkProfile: 'spark3.1' + sparkRuntime: 'spark3.1.3' + - flinkProfile: 'flink1.14' + sparkProfile: 'spark3.0' + sparkRuntime: 'spark3.0.2' + - flinkProfile: 'flink1.14' + sparkProfile: 'spark' + sparkRuntime: 'spark2.4.8' + - flinkProfile: 'flink1.14' + sparkProfile: 'spark2.4' + sparkRuntime: 'spark2.4.8' + steps: + - uses: actions/checkout@v3 + - name: Set up JDK 8 + uses: actions/setup-java@v3 + with: + java-version: '8' + distribution: 'adopt' + architecture: x64 + cache: maven + - name: IT - Bundle Validation - OpenJDK 8 + env: + FLINK_PROFILE: ${{ matrix.flinkProfile }} + SPARK_PROFILE: ${{ matrix.sparkProfile }} + SPARK_RUNTIME: ${{ matrix.sparkRuntime }} + run: | + ./packaging/bundle-validation/ci_run.sh $HUDI_VERSION openjdk8 $STAGING_REPO_NUM + - name: IT - Bundle Validation - OpenJDK 11 + env: + FLINK_PROFILE: ${{ matrix.flinkProfile }} + SPARK_PROFILE: ${{ matrix.sparkProfile }} + SPARK_RUNTIME: ${{ matrix.sparkRuntime }} + if: ${{ startsWith(env.SPARK_PROFILE, 'spark3') }} # Only Spark 3.x supports Java 11 as of now + run: | + ./packaging/bundle-validation/ci_run.sh $HUDI_VERSION openjdk11 $STAGING_REPO_NUM + - name: IT - Bundle Validation - OpenJDK 17 + env: + FLINK_PROFILE: ${{ matrix.flinkProfile }} + SPARK_PROFILE: ${{ matrix.sparkProfile }} + SPARK_RUNTIME: ${{ matrix.sparkRuntime }} + if: ${{ endsWith(env.SPARK_PROFILE, '3.3') }} # Only Spark 3.3 supports Java 17 as of now + run: | + ./packaging/bundle-validation/ci_run.sh $HUDI_VERSION openjdk17 $STAGING_REPO_NUM
