This is an automated email from the ASF dual-hosted git repository.
agrove pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/datafusion-comet.git
The following commit(s) were added to refs/heads/main by this push:
new f4bde1c40 chore: Change default Spark version to 3.5 (#1620)
f4bde1c40 is described below
commit f4bde1c4070f6ff619e88d62eded397c7926596f
Author: Andy Grove <[email protected]>
AuthorDate: Thu Apr 10 08:26:13 2025 -0600
chore: Change default Spark version to 3.5 (#1620)
---
.github/actions/setup-spark-builder/action.yaml | 6 ++----
.github/workflows/docker-publish.yml | 2 +-
docs/source/contributor-guide/debugging.md | 2 +-
docs/source/contributor-guide/development.md | 6 +++---
docs/source/user-guide/datasources.md | 4 ++--
docs/source/user-guide/installation.md | 4 ++--
docs/source/user-guide/source.md | 8 ++++----
kube/Dockerfile | 8 ++++----
pom.xml | 13 ++++++++-----
9 files changed, 27 insertions(+), 26 deletions(-)
diff --git a/.github/actions/setup-spark-builder/action.yaml
b/.github/actions/setup-spark-builder/action.yaml
index dc333e5d1..2b867bada 100644
--- a/.github/actions/setup-spark-builder/action.yaml
+++ b/.github/actions/setup-spark-builder/action.yaml
@@ -19,13 +19,11 @@ name: Setup Spark Builder
description: 'Setup Apache Spark to run SQL tests'
inputs:
spark-short-version:
- description: 'The Apache Spark short version (e.g., 3.4) to build'
+ description: 'The Apache Spark short version (e.g., 3.5) to build'
required: true
- default: '3.4'
spark-version:
- description: 'The Apache Spark version (e.g., 3.4.3) to build'
+ description: 'The Apache Spark version (e.g., 3.5.5) to build'
required: true
- default: '3.4.3'
runs:
using: "composite"
steps:
diff --git a/.github/workflows/docker-publish.yml
b/.github/workflows/docker-publish.yml
index 39741d688..e9af6ac52 100644
--- a/.github/workflows/docker-publish.yml
+++ b/.github/workflows/docker-publish.yml
@@ -73,6 +73,6 @@ jobs:
with:
platforms: linux/amd64,linux/arm64
push: true
- tags: ghcr.io/apache/datafusion-comet:spark-3.4-scala-2.12-${{
env.COMET_VERSION }}
+ tags: ghcr.io/apache/datafusion-comet:spark-3.5-scala-2.12-${{
env.COMET_VERSION }}
file: kube/Dockerfile
no-cache: true
diff --git a/docs/source/contributor-guide/debugging.md
b/docs/source/contributor-guide/debugging.md
index 2f958ce4b..94bdb072b 100644
--- a/docs/source/contributor-guide/debugging.md
+++ b/docs/source/contributor-guide/debugging.md
@@ -130,7 +130,7 @@ Then build the Comet as
[described](https://github.com/apache/arrow-datafusion-c
Start Comet with `RUST_BACKTRACE=1`
```console
-RUST_BACKTRACE=1 $SPARK_HOME/spark-shell --jars
spark/target/comet-spark-spark3.4_2.12-0.8.0-SNAPSHOT.jar --conf
spark.plugins=org.apache.spark.CometPlugin --conf spark.comet.enabled=true
--conf spark.comet.exec.enabled=true
+RUST_BACKTRACE=1 $SPARK_HOME/spark-shell --jars
spark/target/comet-spark-spark3.5_2.12-0.8.0-SNAPSHOT.jar --conf
spark.plugins=org.apache.spark.CometPlugin --conf spark.comet.enabled=true
--conf spark.comet.exec.enabled=true
```
Get the expanded exception details
diff --git a/docs/source/contributor-guide/development.md
b/docs/source/contributor-guide/development.md
index a9c43103b..93d2c709d 100644
--- a/docs/source/contributor-guide/development.md
+++ b/docs/source/contributor-guide/development.md
@@ -109,7 +109,7 @@ The tests can be run with:
```sh
export SPARK_HOME=`pwd`
-./mvnw -pl spark
-Dsuites="org.apache.spark.sql.comet.CometTPCDSV1_4_PlanStabilitySuite" -nsu
test
+./mvnw -pl spark
-Dsuites="org.apache.spark.sql.comet.CometTPCDSV1_4_PlanStabilitySuite"
-Pspark-3.4 -nsu test
./mvnw -pl spark
-Dsuites="org.apache.spark.sql.comet.CometTPCDSV1_4_PlanStabilitySuite"
-Pspark-3.5 -nsu test
./mvnw -pl spark
-Dsuites="org.apache.spark.sql.comet.CometTPCDSV1_4_PlanStabilitySuite"
-Pspark-4.0 -nsu test
```
@@ -117,7 +117,7 @@ export SPARK_HOME=`pwd`
and
```sh
export SPARK_HOME=`pwd`
-./mvnw -pl spark
-Dsuites="org.apache.spark.sql.comet.CometTPCDSV2_7_PlanStabilitySuite" -nsu
test
+./mvnw -pl spark
-Dsuites="org.apache.spark.sql.comet.CometTPCDSV2_7_PlanStabilitySuite"
-Pspark-3.4 -nsu test
./mvnw -pl spark
-Dsuites="org.apache.spark.sql.comet.CometTPCDSV2_7_PlanStabilitySuite"
-Pspark-3.5 -nsu test
./mvnw -pl spark
-Dsuites="org.apache.spark.sql.comet.CometTPCDSV2_7_PlanStabilitySuite"
-Pspark-4.0 -nsu test
```
@@ -127,7 +127,7 @@ To regenerate the golden files, you can run the following
commands.
```sh
export SPARK_HOME=`pwd`
-SPARK_GENERATE_GOLDEN_FILES=1 ./mvnw -pl spark
-Dsuites="org.apache.spark.sql.comet.CometTPCDSV1_4_PlanStabilitySuite" -nsu
test
+SPARK_GENERATE_GOLDEN_FILES=1 ./mvnw -pl spark
-Dsuites="org.apache.spark.sql.comet.CometTPCDSV1_4_PlanStabilitySuite"
-Pspark-3.4 -nsu test
SPARK_GENERATE_GOLDEN_FILES=1 ./mvnw -pl spark
-Dsuites="org.apache.spark.sql.comet.CometTPCDSV1_4_PlanStabilitySuite"
-Pspark-3.5 -nsu test
SPARK_GENERATE_GOLDEN_FILES=1 ./mvnw -pl spark
-Dsuites="org.apache.spark.sql.comet.CometTPCDSV1_4_PlanStabilitySuite"
-Pspark-4.0 -nsu test
```
diff --git a/docs/source/user-guide/datasources.md
b/docs/source/user-guide/datasources.md
index 5cca4b505..5634df8e0 100644
--- a/docs/source/user-guide/datasources.md
+++ b/docs/source/user-guide/datasources.md
@@ -51,12 +51,12 @@ Unlike to native Comet reader the Datafusion reader fully
supports nested types
To build Comet with native DataFusion reader and remote HDFS support it is
required to have a JDK installed
Example:
-Build a Comet for `spark-3.4` provide a JDK path in `JAVA_HOME`
+Build a Comet for `spark-3.5` provide a JDK path in `JAVA_HOME`
Provide the JRE linker path in `RUSTFLAGS`, the path can vary depending on the
system. Typically JRE linker is a part of installed JDK
```shell
export JAVA_HOME="/opt/homebrew/opt/openjdk@11"
-make release PROFILES="-Pspark-3.4" COMET_FEATURES=hdfs RUSTFLAGS="-L
$JAVA_HOME/libexec/openjdk.jdk/Contents/Home/lib/server"
+make release PROFILES="-Pspark-3.5" COMET_FEATURES=hdfs RUSTFLAGS="-L
$JAVA_HOME/libexec/openjdk.jdk/Contents/Home/lib/server"
```
Start Comet with experimental reader and HDFS support as
[described](installation.md/#run-spark-shell-with-comet-enabled)
diff --git a/docs/source/user-guide/installation.md
b/docs/source/user-guide/installation.md
index d42b8de9a..74488bc60 100644
--- a/docs/source/user-guide/installation.md
+++ b/docs/source/user-guide/installation.md
@@ -85,7 +85,7 @@ See the [Comet Kubernetes Guide](kubernetes.md) guide.
Make sure `SPARK_HOME` points to the same Spark version as Comet was built for.
```shell
-export COMET_JAR=spark/target/comet-spark-spark3.4_2.12-0.8.0-SNAPSHOT.jar
+export COMET_JAR=spark/target/comet-spark-spark3.5_2.12-0.8.0-SNAPSHOT.jar
$SPARK_HOME/bin/spark-shell \
--jars $COMET_JAR \
@@ -141,7 +141,7 @@ explicitly contain Comet otherwise Spark may use a
different class-loader for th
components which will then fail at runtime. For example:
```
---driver-class-path spark/target/comet-spark-spark3.4_2.12-0.8.0-SNAPSHOT.jar
+--driver-class-path spark/target/comet-spark-spark3.5_2.12-0.8.0-SNAPSHOT.jar
```
Some cluster managers may require additional configuration, see
<https://spark.apache.org/docs/latest/cluster-overview.html>
diff --git a/docs/source/user-guide/source.md b/docs/source/user-guide/source.md
index ab7f89745..b7038d341 100644
--- a/docs/source/user-guide/source.md
+++ b/docs/source/user-guide/source.md
@@ -38,7 +38,7 @@ cd apache-datafusion-comet-$COMET_VERSION
Build
```console
-make release-nogit PROFILES="-Pspark-3.4"
+make release-nogit PROFILES="-Pspark-3.5"
```
## Building from the GitHub repository
@@ -53,17 +53,17 @@ Build Comet for a specific Spark version:
```console
cd datafusion-comet
-make release PROFILES="-Pspark-3.4"
+make release PROFILES="-Pspark-3.5"
```
Note that the project builds for Scala 2.12 by default but can be built for
Scala 2.13 using an additional profile:
```console
-make release PROFILES="-Pspark-3.4 -Pscala-2.13"
+make release PROFILES="-Pspark-3.5 -Pscala-2.13"
```
To build Comet from the source distribution on an isolated environment without
an access to `github.com` it is necessary to disable
`git-commit-id-maven-plugin`, otherwise you will face errors that there is no
access to the git during the build process. In that case you may use:
```console
-make release-nogit PROFILES="-Pspark-3.4"
+make release-nogit PROFILES="-Pspark-3.5"
```
diff --git a/kube/Dockerfile b/kube/Dockerfile
index 00a526c33..74e087e18 100644
--- a/kube/Dockerfile
+++ b/kube/Dockerfile
@@ -15,7 +15,7 @@
# limitations under the License.
#
-FROM apache/spark:3.4.3 AS builder
+FROM apache/spark:3.5.5 AS builder
USER root
@@ -28,7 +28,7 @@ RUN apt update \
RUN curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh -s -- -y
ENV PATH="/root/.cargo/bin:${PATH}"
ENV RUSTFLAGS="-C debuginfo=line-tables-only -C incremental=false"
-ENV SPARK_VERSION=3.4
+ENV SPARK_VERSION=3.5
ENV SCALA_VERSION=2.12
# copy source files to Docker image
@@ -61,8 +61,8 @@ RUN mkdir -p /root/.m2 && \
RUN cd /comet \
&& JAVA_HOME=$(readlink -f $(which javac) | sed "s/\/bin\/javac//") make
release-nogit PROFILES="-Pspark-$SPARK_VERSION -Pscala-$SCALA_VERSION"
-FROM apache/spark:3.4.3
-ENV SPARK_VERSION=3.4
+FROM apache/spark:3.5.5
+ENV SPARK_VERSION=3.5
ENV SCALA_VERSION=2.12
USER root
diff --git a/pom.xml b/pom.xml
index 9c0339f16..7bdfc782a 100644
--- a/pom.xml
+++ b/pom.xml
@@ -47,13 +47,13 @@ under the License.
<java.version>11</java.version>
<maven.compiler.source>${java.version}</maven.compiler.source>
<maven.compiler.target>${java.version}</maven.compiler.target>
- <scala.version>2.12.17</scala.version>
+ <scala.version>2.12.18</scala.version>
<scala.binary.version>2.12</scala.binary.version>
<scala.plugin.version>4.7.2</scala.plugin.version>
<scalatest.version>3.2.16</scalatest.version>
<scalatest-maven-plugin.version>2.2.0</scalatest-maven-plugin.version>
- <spark.version>3.4.3</spark.version>
- <spark.version.short>3.4</spark.version.short>
+ <spark.version>3.5.5</spark.version>
+ <spark.version.short>3.5</spark.version.short>
<spark.maven.scope>provided</spark.maven.scope>
<protobuf.version>3.25.5</protobuf.version>
<parquet.version>1.13.1</parquet.version>
@@ -64,7 +64,7 @@ under the License.
<spotless.version>2.43.0</spotless.version>
<jacoco.version>0.8.11</jacoco.version>
<semanticdb.version>4.8.8</semanticdb.version>
- <slf4j.version>2.0.6</slf4j.version>
+ <slf4j.version>2.0.7</slf4j.version>
<guava.version>33.2.1-jre</guava.version>
<jni.dir>${project.basedir}/../native/target/debug</jni.dir>
<platform>darwin</platform>
@@ -97,7 +97,7 @@ under the License.
</extraJavaTestArgs>
<argLine>-ea -Xmx4g -Xss4m ${extraJavaTestArgs}</argLine>
<shims.majorVerSrc>spark-3.x</shims.majorVerSrc>
- <shims.minorVerSrc>spark-3.4</shims.minorVerSrc>
+ <shims.minorVerSrc>spark-3.5</shims.minorVerSrc>
</properties>
<dependencyManagement>
@@ -555,8 +555,11 @@ under the License.
<id>spark-3.4</id>
<properties>
<scala.version>2.12.17</scala.version>
+ <spark.version>3.4.3</spark.version>
<spark.version.short>3.4</spark.version.short>
<parquet.version>1.13.1</parquet.version>
+ <slf4j.version>2.0.6</slf4j.version>
+ <shims.minorVerSrc>spark-3.4</shims.minorVerSrc>
</properties>
</profile>
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]