This is an automated email from the ASF dual-hosted git repository. yihua pushed a commit to branch release-1.1.0 in repository https://gitbox.apache.org/repos/asf/hudi.git
commit 5e58dccdd963bea958784077c3f0eadc9e910a6f Author: Y Ethan Guo <[email protected]> AuthorDate: Tue Nov 11 11:22:23 2025 -0500 Improve staging release scripts --- scripts/release/deploy_staging_jars.sh | 17 ++++++++--------- ...ing_jars_java17.sh => deploy_staging_jars_java11.sh} | 7 ++----- scripts/release/deploy_staging_jars_java17.sh | 2 +- 3 files changed, 11 insertions(+), 15 deletions(-) diff --git a/scripts/release/deploy_staging_jars.sh b/scripts/release/deploy_staging_jars.sh index 29aab8168bd4..5906ce65f60f 100755 --- a/scripts/release/deploy_staging_jars.sh +++ b/scripts/release/deploy_staging_jars.sh @@ -46,27 +46,26 @@ declare -a ALL_VERSION_OPTS=( # hudi-utilities-bundle_2.13 # hudi-utilities-slim-bundle_2.13 # hudi-cli-bundle_2.13 -"-Dscala-2.13 -Dspark3.5 -pl hudi-spark-datasource/hudi-spark-common,hudi-spark-datasource/hudi-spark3.5.x,hudi-spark-datasource/hudi-spark,hudi-utilities,packaging/hudi-spark-bundle,packaging/hudi-utilities-bundle,packaging/hudi-utilities-slim-bundle,packaging/hudi-cli-bundle -am" +"-T 1C -Dscala-2.13 -Dspark3.5 -pl hudi-spark-datasource/hudi-spark-common,hudi-spark-datasource/hudi-spark3.5.x,hudi-spark-datasource/hudi-spark,hudi-utilities,packaging/hudi-spark-bundle,packaging/hudi-utilities-bundle,packaging/hudi-utilities-slim-bundle,packaging/hudi-cli-bundle -am" # For Spark 3.3, Scala 2.12: # hudi-spark3.3.x_2.12 # hudi-spark3.3-bundle_2.12 -"-Dscala-2.12 -Dspark3.3 -pl hudi-spark-datasource/hudi-spark3.3.x,packaging/hudi-spark-bundle -am" +"-T 1C -Dscala-2.12 -Dspark3.3 -pl hudi-spark-datasource/hudi-spark3.3.x,packaging/hudi-spark-bundle -am" # For Spark 3.4, Scala 2.12: # hudi-spark3.4.x_2.12 # hudi-spark3.4-bundle_2.12 -"-Dscala-2.12 -Dspark3.4 -pl hudi-spark-datasource/hudi-spark3.4.x,packaging/hudi-spark-bundle -am" +"-T 1C -Dscala-2.12 -Dspark3.4 -pl hudi-spark-datasource/hudi-spark3.4.x,packaging/hudi-spark-bundle -am" # For all modules spark3.5 "-Dscala-2.12 -Dspark3.5" # Upload legacy Spark bundles (not overwriting previous uploads as these jar names are unique) -"-Dscala-2.12 -Dspark3 -pl packaging/hudi-spark-bundle -am" # for legacy bundle name hudi-spark3-bundle_2.12 +"-T 1C -Dscala-2.12 -Dspark3 -pl packaging/hudi-spark-bundle -am" # for legacy bundle name hudi-spark3-bundle_2.12 # Upload Flink bundles (overwriting previous uploads) -"-Dscala-2.12 -Dflink1.17 -Davro.version=1.11.1 -pl packaging/hudi-flink-bundle -am" -"-Dscala-2.12 -Dflink1.18 -Davro.version=1.11.1 -pl packaging/hudi-flink-bundle -am" -"-Dscala-2.12 -Dflink1.19 -Davro.version=1.11.1 -pl packaging/hudi-flink-bundle -am" -"-Dscala-2.12 -Dflink1.20 -Davro.version=1.11.3 -pl packaging/hudi-flink-bundle -am" -"-Dscala-2.12 -Dflink2.0 -Davro.version=1.11.4 -pl packaging/hudi-flink-bundle -am" +"-T 1C -Dscala-2.12 -Dflink1.17 -Davro.version=1.11.4 -Dparquet.version=1.12.3 -pl packaging/hudi-flink-bundle -am" +"-T 1C -Dscala-2.12 -Dflink1.18 -Davro.version=1.11.4 -Dparquet.version=1.13.1 -pl packaging/hudi-flink-bundle -am" +"-T 1C -Dscala-2.12 -Dflink1.19 -Davro.version=1.11.4 -Dparquet.version=1.13.1 -pl packaging/hudi-flink-bundle -am" +"-T 1C -Dscala-2.12 -Dflink1.20 -Davro.version=1.11.4 -Dparquet.version=1.13.1 -pl packaging/hudi-flink-bundle -am" ) printf -v joined "'%s'\n" "${ALL_VERSION_OPTS[@]}" diff --git a/scripts/release/deploy_staging_jars_java17.sh b/scripts/release/deploy_staging_jars_java11.sh similarity index 89% copy from scripts/release/deploy_staging_jars_java17.sh copy to scripts/release/deploy_staging_jars_java11.sh index 02cd8984e9d1..18d15df02256 100755 --- a/scripts/release/deploy_staging_jars_java17.sh +++ b/scripts/release/deploy_staging_jars_java11.sh @@ -37,11 +37,7 @@ if [ "$#" -gt "1" ]; then fi declare -a ALL_VERSION_OPTS=( -# For Spark 4.0, Scala 2.13: -# hudi-spark4-common -# hudi-spark4.0.x_2.13 -# hudi-spark4.0-bundle_2.13 -"-Djava17 -Djava.version=17 -Dscala-2.13 -Dspark4.0 -pl hudi-spark-datasource/hudi-spark4-common,hudi-spark-datasource/hudi-spark4.0.x,packaging/hudi-spark-bundle -am" +"-T 1C -Dscala-2.12 -Dflink2.0 -Davro.version=1.11.4 -Dparquet.version=1.14.4 -pl packaging/hudi-flink-bundle -am" ) printf -v joined "'%s'\n" "${ALL_VERSION_OPTS[@]}" @@ -74,6 +70,7 @@ fi COMMON_OPTIONS="-DdeployArtifacts=true -DskipTests -DretryFailedDeploymentCount=10" for v in "${ALL_VERSION_OPTS[@]}" do + # TODO: consider cleaning all modules by listing directories instead of specifying profile echo "Cleaning everything before any deployment" $MVN clean $COMMON_OPTIONS ${v} echo "Building with options ${v}" diff --git a/scripts/release/deploy_staging_jars_java17.sh b/scripts/release/deploy_staging_jars_java17.sh index 02cd8984e9d1..fbc7aced3a34 100755 --- a/scripts/release/deploy_staging_jars_java17.sh +++ b/scripts/release/deploy_staging_jars_java17.sh @@ -41,7 +41,7 @@ declare -a ALL_VERSION_OPTS=( # hudi-spark4-common # hudi-spark4.0.x_2.13 # hudi-spark4.0-bundle_2.13 -"-Djava17 -Djava.version=17 -Dscala-2.13 -Dspark4.0 -pl hudi-spark-datasource/hudi-spark4-common,hudi-spark-datasource/hudi-spark4.0.x,packaging/hudi-spark-bundle -am" +"-T 1C -Djava17 -Djava.version=17 -Dscala-2.13 -Dspark4.0 -pl hudi-spark-datasource/hudi-spark4-common,hudi-spark-datasource/hudi-spark4.0.x,packaging/hudi-spark-bundle -am" ) printf -v joined "'%s'\n" "${ALL_VERSION_OPTS[@]}"
