This is an automated email from the ASF dual-hosted git repository.

yihua pushed a commit to branch release-1.1.0
in repository https://gitbox.apache.org/repos/asf/hudi.git

commit 5e58dccdd963bea958784077c3f0eadc9e910a6f
Author: Y Ethan Guo <[email protected]>
AuthorDate: Tue Nov 11 11:22:23 2025 -0500

    Improve staging release scripts
---
 scripts/release/deploy_staging_jars.sh                  | 17 ++++++++---------
 ...ing_jars_java17.sh => deploy_staging_jars_java11.sh} |  7 ++-----
 scripts/release/deploy_staging_jars_java17.sh           |  2 +-
 3 files changed, 11 insertions(+), 15 deletions(-)

diff --git a/scripts/release/deploy_staging_jars.sh 
b/scripts/release/deploy_staging_jars.sh
index 29aab8168bd4..5906ce65f60f 100755
--- a/scripts/release/deploy_staging_jars.sh
+++ b/scripts/release/deploy_staging_jars.sh
@@ -46,27 +46,26 @@ declare -a ALL_VERSION_OPTS=(
 # hudi-utilities-bundle_2.13
 # hudi-utilities-slim-bundle_2.13
 # hudi-cli-bundle_2.13
-"-Dscala-2.13 -Dspark3.5 -pl 
hudi-spark-datasource/hudi-spark-common,hudi-spark-datasource/hudi-spark3.5.x,hudi-spark-datasource/hudi-spark,hudi-utilities,packaging/hudi-spark-bundle,packaging/hudi-utilities-bundle,packaging/hudi-utilities-slim-bundle,packaging/hudi-cli-bundle
 -am"
+"-T 1C -Dscala-2.13 -Dspark3.5 -pl 
hudi-spark-datasource/hudi-spark-common,hudi-spark-datasource/hudi-spark3.5.x,hudi-spark-datasource/hudi-spark,hudi-utilities,packaging/hudi-spark-bundle,packaging/hudi-utilities-bundle,packaging/hudi-utilities-slim-bundle,packaging/hudi-cli-bundle
 -am"
 # For Spark 3.3, Scala 2.12:
 # hudi-spark3.3.x_2.12
 # hudi-spark3.3-bundle_2.12
-"-Dscala-2.12 -Dspark3.3 -pl 
hudi-spark-datasource/hudi-spark3.3.x,packaging/hudi-spark-bundle -am"
+"-T 1C -Dscala-2.12 -Dspark3.3 -pl 
hudi-spark-datasource/hudi-spark3.3.x,packaging/hudi-spark-bundle -am"
 # For Spark 3.4, Scala 2.12:
 # hudi-spark3.4.x_2.12
 # hudi-spark3.4-bundle_2.12
-"-Dscala-2.12 -Dspark3.4 -pl 
hudi-spark-datasource/hudi-spark3.4.x,packaging/hudi-spark-bundle -am"
+"-T 1C -Dscala-2.12 -Dspark3.4 -pl 
hudi-spark-datasource/hudi-spark3.4.x,packaging/hudi-spark-bundle -am"
 # For all modules spark3.5
 "-Dscala-2.12 -Dspark3.5"
 
 # Upload legacy Spark bundles (not overwriting previous uploads as these jar 
names are unique)
-"-Dscala-2.12 -Dspark3 -pl packaging/hudi-spark-bundle -am" # for legacy 
bundle name hudi-spark3-bundle_2.12
+"-T 1C -Dscala-2.12 -Dspark3 -pl packaging/hudi-spark-bundle -am" # for legacy 
bundle name hudi-spark3-bundle_2.12
 
 # Upload Flink bundles (overwriting previous uploads)
-"-Dscala-2.12 -Dflink1.17 -Davro.version=1.11.1 -pl 
packaging/hudi-flink-bundle -am"
-"-Dscala-2.12 -Dflink1.18 -Davro.version=1.11.1 -pl 
packaging/hudi-flink-bundle -am"
-"-Dscala-2.12 -Dflink1.19 -Davro.version=1.11.1 -pl 
packaging/hudi-flink-bundle -am"
-"-Dscala-2.12 -Dflink1.20 -Davro.version=1.11.3 -pl 
packaging/hudi-flink-bundle -am"
-"-Dscala-2.12 -Dflink2.0 -Davro.version=1.11.4 -pl packaging/hudi-flink-bundle 
-am"
+"-T 1C -Dscala-2.12 -Dflink1.17 -Davro.version=1.11.4 -Dparquet.version=1.12.3 
-pl packaging/hudi-flink-bundle -am"
+"-T 1C -Dscala-2.12 -Dflink1.18 -Davro.version=1.11.4 -Dparquet.version=1.13.1 
-pl packaging/hudi-flink-bundle -am"
+"-T 1C -Dscala-2.12 -Dflink1.19 -Davro.version=1.11.4 -Dparquet.version=1.13.1 
-pl packaging/hudi-flink-bundle -am"
+"-T 1C -Dscala-2.12 -Dflink1.20 -Davro.version=1.11.4 -Dparquet.version=1.13.1 
-pl packaging/hudi-flink-bundle -am"
 )
 printf -v joined "'%s'\n" "${ALL_VERSION_OPTS[@]}"
 
diff --git a/scripts/release/deploy_staging_jars_java17.sh 
b/scripts/release/deploy_staging_jars_java11.sh
similarity index 89%
copy from scripts/release/deploy_staging_jars_java17.sh
copy to scripts/release/deploy_staging_jars_java11.sh
index 02cd8984e9d1..18d15df02256 100755
--- a/scripts/release/deploy_staging_jars_java17.sh
+++ b/scripts/release/deploy_staging_jars_java11.sh
@@ -37,11 +37,7 @@ if [ "$#" -gt "1" ]; then
 fi
 
 declare -a ALL_VERSION_OPTS=(
-# For Spark 4.0, Scala 2.13:
-# hudi-spark4-common
-# hudi-spark4.0.x_2.13
-# hudi-spark4.0-bundle_2.13
-"-Djava17 -Djava.version=17 -Dscala-2.13 -Dspark4.0 -pl 
hudi-spark-datasource/hudi-spark4-common,hudi-spark-datasource/hudi-spark4.0.x,packaging/hudi-spark-bundle
 -am"
+"-T 1C -Dscala-2.12 -Dflink2.0 -Davro.version=1.11.4 -Dparquet.version=1.14.4 
-pl packaging/hudi-flink-bundle -am"
 )
 printf -v joined "'%s'\n" "${ALL_VERSION_OPTS[@]}"
 
@@ -74,6 +70,7 @@ fi
 COMMON_OPTIONS="-DdeployArtifacts=true -DskipTests 
-DretryFailedDeploymentCount=10"
 for v in "${ALL_VERSION_OPTS[@]}"
 do
+  # TODO: consider cleaning all modules by listing directories instead of 
specifying profile
   echo "Cleaning everything before any deployment"
   $MVN clean $COMMON_OPTIONS ${v}
   echo "Building with options ${v}"
diff --git a/scripts/release/deploy_staging_jars_java17.sh 
b/scripts/release/deploy_staging_jars_java17.sh
index 02cd8984e9d1..fbc7aced3a34 100755
--- a/scripts/release/deploy_staging_jars_java17.sh
+++ b/scripts/release/deploy_staging_jars_java17.sh
@@ -41,7 +41,7 @@ declare -a ALL_VERSION_OPTS=(
 # hudi-spark4-common
 # hudi-spark4.0.x_2.13
 # hudi-spark4.0-bundle_2.13
-"-Djava17 -Djava.version=17 -Dscala-2.13 -Dspark4.0 -pl 
hudi-spark-datasource/hudi-spark4-common,hudi-spark-datasource/hudi-spark4.0.x,packaging/hudi-spark-bundle
 -am"
+"-T 1C -Djava17 -Djava.version=17 -Dscala-2.13 -Dspark4.0 -pl 
hudi-spark-datasource/hudi-spark4-common,hudi-spark-datasource/hudi-spark4.0.x,packaging/hudi-spark-bundle
 -am"
 )
 printf -v joined "'%s'\n" "${ALL_VERSION_OPTS[@]}"
 

Reply via email to