This is an automated email from the ASF dual-hosted git repository.

yangzy pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/incubator-gluten.git


The following commit(s) were added to refs/heads/main by this push:
     new 94ecd9e25a [GLUTEN-7336][CORE] Bump Spark version to v3.5.3 (#7537)
94ecd9e25a is described below

commit 94ecd9e25af1e036c6431b1873b72efb0a907d6d
Author: Yang Zhang <[email protected]>
AuthorDate: Tue Oct 22 16:53:38 2024 +0800

    [GLUTEN-7336][CORE] Bump Spark version to v3.5.3 (#7537)
---
 .github/workflows/util/install_spark_resources.sh    | 20 ++++++++++----------
 .github/workflows/velox_backend.yml                  | 14 +++++++-------
 pom.xml                                              |  8 ++++----
 .../gluten/sql/shims/spark35/SparkShimProvider.scala |  2 +-
 tools/gluten-it/pom.xml                              |  2 +-
 5 files changed, 23 insertions(+), 23 deletions(-)

diff --git a/.github/workflows/util/install_spark_resources.sh 
b/.github/workflows/util/install_spark_resources.sh
index dd2afec821..d5f413ee21 100755
--- a/.github/workflows/util/install_spark_resources.sh
+++ b/.github/workflows/util/install_spark_resources.sh
@@ -63,26 +63,26 @@ case "$1" in
 3.5)
     # Spark-3.5
     cd ${INSTALL_DIR} && \
-    wget -nv 
https://archive.apache.org/dist/spark/spark-3.5.2/spark-3.5.2-bin-hadoop3.tgz 
&& \
-    tar --strip-components=1 -xf spark-3.5.2-bin-hadoop3.tgz 
spark-3.5.2-bin-hadoop3/jars/ && \
-    rm -rf spark-3.5.2-bin-hadoop3.tgz && \
+    wget -nv 
https://archive.apache.org/dist/spark/spark-3.5.3/spark-3.5.3-bin-hadoop3.tgz 
&& \
+    tar --strip-components=1 -xf spark-3.5.3-bin-hadoop3.tgz 
spark-3.5.3-bin-hadoop3/jars/ && \
+    rm -rf spark-3.5.3-bin-hadoop3.tgz && \
     mkdir -p 
${INSTALL_DIR}/shims/spark35/spark_home/assembly/target/scala-2.12 && \
     mv jars ${INSTALL_DIR}/shims/spark35/spark_home/assembly/target/scala-2.12 
&& \
-    wget -nv https://github.com/apache/spark/archive/refs/tags/v3.5.2.tar.gz 
&& \
-    tar --strip-components=1 -xf v3.5.2.tar.gz 
spark-3.5.2/sql/core/src/test/resources/  && \
+    wget -nv https://github.com/apache/spark/archive/refs/tags/v3.5.3.tar.gz 
&& \
+    tar --strip-components=1 -xf v3.5.3.tar.gz 
spark-3.5.3/sql/core/src/test/resources/  && \
     mkdir -p shims/spark35/spark_home/ && \
     mv sql shims/spark35/spark_home/
     ;;
 3.5-scala2.13)
     # Spark-3.5, scala 2.13
     cd ${INSTALL_DIR} && \
-    wget -nv 
https://archive.apache.org/dist/spark/spark-3.5.2/spark-3.5.2-bin-hadoop3.tgz 
&& \
-    tar --strip-components=1 -xf spark-3.5.2-bin-hadoop3.tgz 
spark-3.5.2-bin-hadoop3/jars/ && \
-    rm -rf spark-3.5.2-bin-hadoop3.tgz && \
+    wget -nv 
https://archive.apache.org/dist/spark/spark-3.5.3/spark-3.5.3-bin-hadoop3.tgz 
&& \
+    tar --strip-components=1 -xf spark-3.5.3-bin-hadoop3.tgz 
spark-3.5.3-bin-hadoop3/jars/ && \
+    rm -rf spark-3.5.3-bin-hadoop3.tgz && \
     mkdir -p 
${INSTALL_DIR}/shims/spark35/spark_home/assembly/target/scala-2.13 && \
     mv jars ${INSTALL_DIR}/shims/spark35/spark_home/assembly/target/scala-2.13 
&& \
-    wget -nv https://github.com/apache/spark/archive/refs/tags/v3.5.2.tar.gz 
&& \
-    tar --strip-components=1 -xf v3.5.2.tar.gz 
spark-3.5.2/sql/core/src/test/resources/  && \
+    wget -nv https://github.com/apache/spark/archive/refs/tags/v3.5.3.tar.gz 
&& \
+    tar --strip-components=1 -xf v3.5.3.tar.gz 
spark-3.5.3/sql/core/src/test/resources/  && \
     mkdir -p shims/spark35/spark_home/ && \
     mv sql shims/spark35/spark_home/
     ;;
diff --git a/.github/workflows/velox_backend.yml 
b/.github/workflows/velox_backend.yml
index 8637c92ffe..dd31b8e77b 100644
--- a/.github/workflows/velox_backend.yml
+++ b/.github/workflows/velox_backend.yml
@@ -860,13 +860,13 @@ jobs:
           dnf module -y install python39 && \
           alternatives --set python3 /usr/bin/python3.9 && \
           pip3 install setuptools && \
-          pip3 install pyspark==3.5.2 cython && \
+          pip3 install pyspark==3.5.3 cython && \
           pip3 install pandas pyarrow
-      - name: Build and Run unit test for Spark 3.5.2 (other tests)
+      - name: Build and Run unit test for Spark 3.5.3 (other tests)
         run: |
           cd $GITHUB_WORKSPACE/
           export SPARK_SCALA_VERSION=2.12
-          $MVN_CMD clean test -Pspark-3.5 -Pbackends-velox -Pceleborn 
-Piceberg -Pdelta -Phudi -Pspark-ut \
+          $MVN_CMD clean test -Pspark-3.5 -Pbackends-velox -Pceleborn 
-Piceberg -Pdelta -Pspark-ut \
           
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark35/spark_home/" \
           
-DtagsToExclude=org.apache.spark.tags.ExtendedSQLTest,org.apache.gluten.tags.UDFTest,org.apache.gluten.tags.SkipTestTags
       - name: Upload test report
@@ -906,9 +906,9 @@ jobs:
           dnf module -y install python39 && \
           alternatives --set python3 /usr/bin/python3.9 && \
           pip3 install setuptools && \
-          pip3 install pyspark==3.5.2 cython && \
+          pip3 install pyspark==3.5.3 cython && \
           pip3 install pandas pyarrow
-      - name: Build and Run unit test for Spark 3.5.2 with scala-2.13 (other 
tests)
+      - name: Build and Run unit test for Spark 3.5.3 with scala-2.13 (other 
tests)
         run: |
           cd $GITHUB_WORKSPACE/
           export SPARK_SCALA_VERSION=2.13
@@ -949,10 +949,10 @@ jobs:
       - name: Prepare spark.test.home for Spark 3.5.2 (other tests)
         run: |
           bash .github/workflows/util/install_spark_resources.sh 3.5
-      - name: Build and Run unit test for Spark 3.5.2 (slow tests)
+      - name: Build and Run unit test for Spark 3.5.3 (slow tests)
         run: |
           cd $GITHUB_WORKSPACE/
-          $MVN_CMD clean test -Pspark-3.5 -Pbackends-velox -Pceleborn 
-Piceberg -Pdelta -Phudi -Pspark-ut \
+          $MVN_CMD clean test -Pspark-3.5 -Pbackends-velox -Pceleborn 
-Piceberg -Pdelta -Pspark-ut \
           
-DargLine="-Dspark.test.home=$GITHUB_WORKSPACE//shims/spark35/spark_home/" \
           -DtagsToInclude=org.apache.spark.tags.ExtendedSQLTest
       - name: Upload test report
diff --git a/pom.xml b/pom.xml
index a3ce1c6f70..ea8a9f9808 100644
--- a/pom.xml
+++ b/pom.xml
@@ -341,12 +341,12 @@
         <sparkbundle.version>3.5</sparkbundle.version>
         <sparkshim.module.name>spark35</sparkshim.module.name>
         
<sparkshim.artifactId>spark-sql-columnar-shims-spark35</sparkshim.artifactId>
-        <spark.version>3.5.2</spark.version>
+        <spark.version>3.5.3</spark.version>
         <iceberg.version>1.5.0</iceberg.version>
         <delta.package.name>delta-spark</delta.package.name>
-        <delta.version>3.2.0</delta.version>
-       <delta.binary.version>32</delta.binary.version>
-       <hudi.version>0.15.0</hudi.version>
+        <delta.version>3.2.1</delta.version>
+        <delta.binary.version>32</delta.binary.version>
+           <hudi.version>0.15.0</hudi.version>
         <fasterxml.version>2.15.1</fasterxml.version>
         <hadoop.version>3.3.4</hadoop.version>
         <antlr4.version>4.9.3</antlr4.version>
diff --git 
a/shims/spark35/src/main/scala/org/apache/gluten/sql/shims/spark35/SparkShimProvider.scala
 
b/shims/spark35/src/main/scala/org/apache/gluten/sql/shims/spark35/SparkShimProvider.scala
index eab32ab9d0..ad99f7be25 100644
--- 
a/shims/spark35/src/main/scala/org/apache/gluten/sql/shims/spark35/SparkShimProvider.scala
+++ 
b/shims/spark35/src/main/scala/org/apache/gluten/sql/shims/spark35/SparkShimProvider.scala
@@ -20,7 +20,7 @@ import org.apache.gluten.sql.shims.{SparkShimDescriptor, 
SparkShims}
 import org.apache.gluten.sql.shims.spark35.SparkShimProvider.DESCRIPTOR
 
 object SparkShimProvider {
-  val DESCRIPTOR = SparkShimDescriptor(3, 5, 2)
+  val DESCRIPTOR = SparkShimDescriptor(3, 5, 3)
 }
 
 class SparkShimProvider extends org.apache.gluten.sql.shims.SparkShimProvider {
diff --git a/tools/gluten-it/pom.xml b/tools/gluten-it/pom.xml
index bad4d6087f..70b8d20c5b 100644
--- a/tools/gluten-it/pom.xml
+++ b/tools/gluten-it/pom.xml
@@ -164,7 +164,7 @@
     <profile>
       <id>spark-3.5</id>
       <properties>
-        <spark.version>3.5.2</spark.version>
+        <spark.version>3.5.3</spark.version>
         <scala.library.version>2.12.18</scala.library.version>
       </properties>
     </profile>


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to