(spark) branch master updated (081809667611 -> c4e4497ff7e7)

2024-02-18 Thread weichenxu123
This is an automated email from the ASF dual-hosted git repository.

weichenxu123 pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


from 081809667611 [MINOR][SQL] Remove `unsupportedOperationMsg` from 
`CaseInsensitiveStringMap`
 add c4e4497ff7e7 [SPARK-46812][SQL][PYTHON] Make mapInPandas / mapInArrow 
support ResourceProfile

No new revisions were added by this update.

Summary of changes:
 .../sql/connect/planner/SparkConnectPlanner.scala  |   6 +-
 dev/sparktestsupport/modules.py|   1 +
 python/pyspark/sql/pandas/map_ops.py   |  61 +++-
 python/pyspark/sql/tests/test_resources.py | 104 +
 .../catalyst/analysis/DeduplicateRelations.scala   |   4 +-
 .../plans/logical/pythonLogicalOperators.scala |   7 +-
 .../sql/catalyst/analysis/AnalysisSuite.scala  |   9 +-
 .../main/scala/org/apache/spark/sql/Dataset.scala  |  17 +++-
 .../spark/sql/execution/SparkStrategies.scala  |   8 +-
 .../sql/execution/python/MapInArrowExec.scala  |   4 +-
 .../sql/execution/python/MapInBatchExec.scala  |   6 +-
 .../sql/execution/python/MapInPandasExec.scala |   4 +-
 12 files changed, 206 insertions(+), 25 deletions(-)
 create mode 100644 python/pyspark/sql/tests/test_resources.py


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



(spark) branch branch-3.5 updated: [SPARK-47072][SQL][3.5] Fix supported interval formats in error messages

2024-02-18 Thread maxgekk
This is an automated email from the ASF dual-hosted git repository.

maxgekk pushed a commit to branch branch-3.5
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.5 by this push:
 new 93a09ea279e6 [SPARK-47072][SQL][3.5] Fix supported interval formats in 
error messages
93a09ea279e6 is described below

commit 93a09ea279e6bd2515ced66d8f38053e4a5514ce
Author: Max Gekk 
AuthorDate: Mon Feb 19 10:29:08 2024 +0300

[SPARK-47072][SQL][3.5] Fix supported interval formats in error messages

### What changes were proposed in this pull request?
In the PR, I propose to add one more field to keys of `supportedFormat` in 
`IntervalUtils` because current implementation has duplicate keys that 
overwrites each other. For instance, the following keys are the same:
```
(YM.YEAR, YM.MONTH)
...
(DT.DAY, DT.HOUR)
```
because `YM.YEAR = DT.DAY = 0` and `YM.MONTH = DT.HOUR = 1`.

This is a backport of https://github.com/apache/spark/pull/45127.

### Why are the changes needed?
To fix the incorrect error message when Spark cannot parse ANSI interval 
string. For example, the expected format should be some year-month format but 
Spark outputs day-time one:
```sql
spark-sql (default)> select interval '-\t2-2\t' year to month;

Interval string does not match year-month format of `[+|-]d h`, `INTERVAL 
[+|-]'[+|-]d h' DAY TO HOUR` when cast to interval year to month: -   2-2 . 
(line 1, pos 16)

== SQL ==
select interval '-\t2-2\t' year to month
^^^
```

### Does this PR introduce _any_ user-facing change?
Yes.

### How was this patch tested?
By running the existing test suite:
```
$ build/sbt "test:testOnly *IntervalUtilsSuite"
```
and regenerating the golden files:
```
$ SPARK_GENERATE_GOLDEN_FILES=1 PYSPARK_PYTHON=python3 build/sbt 
"sql/testOnly org.apache.spark.sql.SQLQueryTestSuite"
```

### Was this patch authored or co-authored using generative AI tooling?
No.

Authored-by: Max Gekk 
(cherry picked from commit 074fcf2807000d342831379de0fafc1e49a6bf19)

Closes #45139 from MaxGekk/fix-supportedFormat-3.5.

Authored-by: Max Gekk 
Signed-off-by: Max Gekk 
---
 .../spark/sql/catalyst/util/IntervalUtils.scala| 33 +-
 .../sql/catalyst/expressions/CastSuiteBase.scala   |  8 +++---
 .../analyzer-results/ansi/interval.sql.out |  2 +-
 .../sql-tests/analyzer-results/interval.sql.out|  2 +-
 .../sql-tests/results/ansi/interval.sql.out|  2 +-
 .../resources/sql-tests/results/interval.sql.out   |  2 +-
 6 files changed, 27 insertions(+), 22 deletions(-)

diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/IntervalUtils.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/IntervalUtils.scala
index e051cfc37f12..4d90007400ea 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/IntervalUtils.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/IntervalUtils.scala
@@ -107,25 +107,30 @@ object IntervalUtils extends SparkIntervalUtils {
   fallBackNotice: Option[String] = None) = {
 throw new IllegalArgumentException(
   s"Interval string does not match $intervalStr format of " +
-s"${supportedFormat((startFiled, endField)).map(format => 
s"`$format`").mkString(", ")} " +
+s"${supportedFormat((intervalStr, startFiled, endField))
+  .map(format => s"`$format`").mkString(", ")} " +
 s"when cast to $typeName: ${input.toString}" +
 s"${fallBackNotice.map(s => s", $s").getOrElse("")}")
   }
 
   val supportedFormat = Map(
-(YM.YEAR, YM.MONTH) -> Seq("[+|-]y-m", "INTERVAL [+|-]'[+|-]y-m' YEAR TO 
MONTH"),
-(YM.YEAR, YM.YEAR) -> Seq("[+|-]y", "INTERVAL [+|-]'[+|-]y' YEAR"),
-(YM.MONTH, YM.MONTH) -> Seq("[+|-]m", "INTERVAL [+|-]'[+|-]m' MONTH"),
-(DT.DAY, DT.DAY) -> Seq("[+|-]d", "INTERVAL [+|-]'[+|-]d' DAY"),
-(DT.DAY, DT.HOUR) -> Seq("[+|-]d h", "INTERVAL [+|-]'[+|-]d h' DAY TO 
HOUR"),
-(DT.DAY, DT.MINUTE) -> Seq("[+|-]d h:m", "INTERVAL [+|-]'[+|-]d h:m' DAY 
TO MINUTE"),
-(DT.DAY, DT.SECOND) -> Seq("[+|-]d h:m:s.n", "INTERVAL [+|-]'[+|-]d 
h:m:s.n' DAY TO SECOND"),
-(DT.HOUR, DT.HOUR) -> Seq("[+|-]h", "INTERVAL [+|-]'[+|-]h' HOUR"),
-(DT.HOUR, DT.MINUTE) -> Seq("[+|-]h:m", "INTERVAL [+|-]'[+|-]h:m' HOUR TO 
MINUTE"),
-(DT.HOUR, DT.SECOND) -> Seq("[+|-]h:m:s.n", "INTERVAL [+|-]'[+|-]h:m:s.n' 
HOUR TO SECOND"),
-(DT.MINUTE, DT.MINUTE) -> Seq("[+|-]m", "INTERVAL [+|-]'[+|-]m' MINUTE"),
-(DT.MINUTE, DT.SECOND) -> Seq("[+|-]m:s.n", "INTERVAL [+|-]'[+|-]m:s.n' 
MINUTE TO SECOND"),
-(DT.SECOND, DT.SECOND) -> Seq("[+|-]s.n", "INTERVAL [+|-]'[+|-]s.n' 
SECOND")
+("year-month", YM.YEAR, YM.MONTH) -> Seq("[+|-]y-m", "INTERVAL 

(spark) branch branch-3.4 updated: [SPARK-47072][SQL][3.4] Fix supported interval formats in error messages

2024-02-18 Thread maxgekk
This is an automated email from the ASF dual-hosted git repository.

maxgekk pushed a commit to branch branch-3.4
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/branch-3.4 by this push:
 new 081c7a7947a4 [SPARK-47072][SQL][3.4] Fix supported interval formats in 
error messages
081c7a7947a4 is described below

commit 081c7a7947a47bf0b2bfd478abdd4b78a1db3ddb
Author: Max Gekk 
AuthorDate: Mon Feb 19 10:26:47 2024 +0300

[SPARK-47072][SQL][3.4] Fix supported interval formats in error messages

### What changes were proposed in this pull request?
In the PR, I propose to add one more field to keys of `supportedFormat` in 
`IntervalUtils` because current implementation has duplicate keys that 
overwrites each other. For instance, the following keys are the same:
```
(YM.YEAR, YM.MONTH)
...
(DT.DAY, DT.HOUR)
```
because `YM.YEAR = DT.DAY = 0` and `YM.MONTH = DT.HOUR = 1`.

This is a backport of https://github.com/apache/spark/pull/45127.

### Why are the changes needed?
To fix the incorrect error message when Spark cannot parse ANSI interval 
string. For example, the expected format should be some year-month format but 
Spark outputs day-time one:
```sql
spark-sql (default)> select interval '-\t2-2\t' year to month;

Interval string does not match year-month format of `[+|-]d h`, `INTERVAL 
[+|-]'[+|-]d h' DAY TO HOUR` when cast to interval year to month: -   2-2 . 
(line 1, pos 16)

== SQL ==
select interval '-\t2-2\t' year to month
^^^
```

### Does this PR introduce _any_ user-facing change?
Yes.

### How was this patch tested?
By running the existing test suite:
```
$ build/sbt "test:testOnly *IntervalUtilsSuite"
```
and regenerating the golden files:
```
$ SPARK_GENERATE_GOLDEN_FILES=1 PYSPARK_PYTHON=python3 build/sbt 
"sql/testOnly org.apache.spark.sql.SQLQueryTestSuite"
```

### Was this patch authored or co-authored using generative AI tooling?
No.

Authored-by: Max Gekk 
(cherry picked from commit 074fcf2807000d342831379de0fafc1e49a6bf19)

Closes #45140 from MaxGekk/fix-supportedFormat-3.4.

Authored-by: Max Gekk 
Signed-off-by: Max Gekk 
---
 .../spark/sql/catalyst/util/IntervalUtils.scala| 33 +-
 .../sql/catalyst/expressions/CastSuiteBase.scala   |  8 +++---
 .../sql-tests/results/ansi/interval.sql.out|  2 +-
 .../resources/sql-tests/results/interval.sql.out   |  2 +-
 4 files changed, 25 insertions(+), 20 deletions(-)

diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/IntervalUtils.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/IntervalUtils.scala
index 7b574e987d91..93b8f2f512ea 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/IntervalUtils.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/util/IntervalUtils.scala
@@ -124,25 +124,30 @@ object IntervalUtils {
   fallBackNotice: Option[String] = None) = {
 throw new IllegalArgumentException(
   s"Interval string does not match $intervalStr format of " +
-s"${supportedFormat((startFiled, endField)).map(format => 
s"`$format`").mkString(", ")} " +
+s"${supportedFormat((intervalStr, startFiled, endField))
+  .map(format => s"`$format`").mkString(", ")} " +
 s"when cast to $typeName: ${input.toString}" +
 s"${fallBackNotice.map(s => s", $s").getOrElse("")}")
   }
 
   val supportedFormat = Map(
-(YM.YEAR, YM.MONTH) -> Seq("[+|-]y-m", "INTERVAL [+|-]'[+|-]y-m' YEAR TO 
MONTH"),
-(YM.YEAR, YM.YEAR) -> Seq("[+|-]y", "INTERVAL [+|-]'[+|-]y' YEAR"),
-(YM.MONTH, YM.MONTH) -> Seq("[+|-]m", "INTERVAL [+|-]'[+|-]m' MONTH"),
-(DT.DAY, DT.DAY) -> Seq("[+|-]d", "INTERVAL [+|-]'[+|-]d' DAY"),
-(DT.DAY, DT.HOUR) -> Seq("[+|-]d h", "INTERVAL [+|-]'[+|-]d h' DAY TO 
HOUR"),
-(DT.DAY, DT.MINUTE) -> Seq("[+|-]d h:m", "INTERVAL [+|-]'[+|-]d h:m' DAY 
TO MINUTE"),
-(DT.DAY, DT.SECOND) -> Seq("[+|-]d h:m:s.n", "INTERVAL [+|-]'[+|-]d 
h:m:s.n' DAY TO SECOND"),
-(DT.HOUR, DT.HOUR) -> Seq("[+|-]h", "INTERVAL [+|-]'[+|-]h' HOUR"),
-(DT.HOUR, DT.MINUTE) -> Seq("[+|-]h:m", "INTERVAL [+|-]'[+|-]h:m' HOUR TO 
MINUTE"),
-(DT.HOUR, DT.SECOND) -> Seq("[+|-]h:m:s.n", "INTERVAL [+|-]'[+|-]h:m:s.n' 
HOUR TO SECOND"),
-(DT.MINUTE, DT.MINUTE) -> Seq("[+|-]m", "INTERVAL [+|-]'[+|-]m' MINUTE"),
-(DT.MINUTE, DT.SECOND) -> Seq("[+|-]m:s.n", "INTERVAL [+|-]'[+|-]m:s.n' 
MINUTE TO SECOND"),
-(DT.SECOND, DT.SECOND) -> Seq("[+|-]s.n", "INTERVAL [+|-]'[+|-]s.n' 
SECOND")
+("year-month", YM.YEAR, YM.MONTH) -> Seq("[+|-]y-m", "INTERVAL 
[+|-]'[+|-]y-m' YEAR TO MONTH"),
+("year-month", YM.YEAR, YM.YEAR) -> Seq("[+|-]y", "INTERVAL [+|-]'[+|-]y' 
YEAR"),
+("year-month", YM.MONTH, 

(spark) branch master updated: [MINOR][SQL] Remove `unsupportedOperationMsg` from `CaseInsensitiveStringMap`

2024-02-18 Thread maxgekk
This is an automated email from the ASF dual-hosted git repository.

maxgekk pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 081809667611 [MINOR][SQL] Remove `unsupportedOperationMsg` from 
`CaseInsensitiveStringMap`
081809667611 is described below

commit 081809667611ceca926e9ef66f0bc3a786039314
Author: yangjie01 
AuthorDate: Mon Feb 19 10:25:01 2024 +0300

[MINOR][SQL] Remove `unsupportedOperationMsg` from 
`CaseInsensitiveStringMap`

### What changes were proposed in this pull request?
This pr remove unused private field `unsupportedOperationMsg` from 
`CaseInsensitiveStringMap`, it is replaced by `_LEGACY_ERROR_TEMP_3206` after 
https://github.com/apache/spark/pull/44549 is merged,

### Why are the changes needed?
Remove unused private field.

### Does this PR introduce _any_ user-facing change?
No

### How was this patch tested?
Pass GitHub Actions

### Was this patch authored or co-authored using generative AI tooling?
No

Closes #45161 from LuciferYang/minior-CaseInsensitiveStringMap.

Authored-by: yangjie01 
Signed-off-by: Max Gekk 
---
 .../main/java/org/apache/spark/sql/util/CaseInsensitiveStringMap.java   | 2 --
 1 file changed, 2 deletions(-)

diff --git 
a/sql/catalyst/src/main/java/org/apache/spark/sql/util/CaseInsensitiveStringMap.java
 
b/sql/catalyst/src/main/java/org/apache/spark/sql/util/CaseInsensitiveStringMap.java
index 1c5c38ba705f..00a3de692fbf 100644
--- 
a/sql/catalyst/src/main/java/org/apache/spark/sql/util/CaseInsensitiveStringMap.java
+++ 
b/sql/catalyst/src/main/java/org/apache/spark/sql/util/CaseInsensitiveStringMap.java
@@ -45,8 +45,6 @@ import java.util.Set;
 public class CaseInsensitiveStringMap implements Map {
   private static final Logger logger = 
LoggerFactory.getLogger(CaseInsensitiveStringMap.class);
 
-  private String unsupportedOperationMsg = "CaseInsensitiveStringMap is 
read-only.";
-
   public static CaseInsensitiveStringMap empty() {
 return new CaseInsensitiveStringMap(new HashMap<>(0));
   }


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



(spark) branch master updated (174a19c1c039 -> 91dfc31ff89c)

2024-02-18 Thread yao
This is an automated email from the ASF dual-hosted git repository.

yao pushed a change to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


from 174a19c1c039 [MINOR][INFRA] Rename build_maven.yml and 
build_maven_java21.yml
 add 91dfc31ff89c [SPARK-47085][SQL] reduce the complexity of toTRowSet 
from n^2 to n

No new revisions were added by this update.

Summary of changes:
 .../spark/sql/hive/thriftserver/RowSetUtils.scala  | 23 +++---
 .../SparkExecuteStatementOperation.scala   |  2 +-
 2 files changed, 8 insertions(+), 17 deletions(-)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



(spark) branch master updated: [MINOR][INFRA] Rename build_maven.yml and build_maven_java21.yml

2024-02-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 174a19c1c039 [MINOR][INFRA] Rename build_maven.yml and 
build_maven_java21.yml
174a19c1c039 is described below

commit 174a19c1c0391b1084e6a0b8fdbc5c019a662b01
Author: Hyukjin Kwon 
AuthorDate: Mon Feb 19 12:45:32 2024 +0900

[MINOR][INFRA] Rename build_maven.yml and build_maven_java21.yml

### What changes were proposed in this pull request?

This PR proposes to change the name of `build_maven.yml` and 
`build_maven_java21.yml` from
- Build using Maven (master, Scala 2.13, Hadoop 3, JDK 17) -> Build \ Maven 
(master, Scala 2.13, Hadoop 3, JDK 17)
- Build using Maven (master, Scala 2.13, Hadoop 3, JDK 21) -> Build \ Maven 
(master, Scala 2.13, Hadoop 3, JDK 21)

### Why are the changes needed?

To make it easier to navigate, and to be consistent. By doing this, you 
don't have to click "Show more workloads" at 
https://github.com/apache/spark/actions

![Screenshot 2024-02-19 at 12 40 51 
PM](https://github.com/apache/spark/assets/6477701/a7cba7c2-1264-40d1-b629-7713d22d6535)

### Does this PR introduce _any_ user-facing change?

No, dev-only.

### How was this patch tested?

Will be tested in this PR's GitHub Actions.

### Was this patch authored or co-authored using generative AI tooling?

No.

Closes #45160 from HyukjinKwon/maven-builds.

Authored-by: Hyukjin Kwon 
Signed-off-by: Hyukjin Kwon 
---
 .github/workflows/build_maven.yml| 2 +-
 .github/workflows/build_maven_java21.yml | 2 +-
 2 files changed, 2 insertions(+), 2 deletions(-)

diff --git a/.github/workflows/build_maven.yml 
b/.github/workflows/build_maven.yml
index d43366fa86e7..863c5e16ba4b 100644
--- a/.github/workflows/build_maven.yml
+++ b/.github/workflows/build_maven.yml
@@ -17,7 +17,7 @@
 # under the License.
 #
 
-name: "Build using Maven (master, Scala 2.13, Hadoop 3, JDK 17)"
+name: "Build / Maven (master, Scala 2.13, Hadoop 3, JDK 17)"
 
 on:
   schedule:
diff --git a/.github/workflows/build_maven_java21.yml 
b/.github/workflows/build_maven_java21.yml
index e50f996a50ba..127904145464 100644
--- a/.github/workflows/build_maven_java21.yml
+++ b/.github/workflows/build_maven_java21.yml
@@ -17,7 +17,7 @@
 # under the License.
 #
 
-name: "Build using Maven (master, Scala 2.13, Hadoop 3, JDK 21)"
+name: "Build / Maven (master, Scala 2.13, Hadoop 3, JDK 21)"
 
 on:
   schedule:


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



(spark) branch master updated: [SPARK-47066][INFRA][FOLLOW-UP] Deduplicate Apple Silicon Maven build definition

2024-02-18 Thread gurwls223
This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 8a9a1be9b02a [SPARK-47066][INFRA][FOLLOW-UP] Deduplicate Apple Silicon 
Maven build definition
8a9a1be9b02a is described below

commit 8a9a1be9b02ad772aa5bdb61429b92a70e894cce
Author: Hyukjin Kwon 
AuthorDate: Mon Feb 19 12:10:46 2024 +0900

[SPARK-47066][INFRA][FOLLOW-UP] Deduplicate Apple Silicon Maven build 
definition

### What changes were proposed in this pull request?

This PR is a followup of https://github.com/apache/spark/pull/45126 that 
deduplicates the build code.

### Why are the changes needed?

For better maintenance of the code

### Does this PR introduce _any_ user-facing change?

No, dev-only.

### How was this patch tested?

This PR should test it out.

### Was this patch authored or co-authored using generative AI tooling?

No.

Closes #45148 from HyukjinKwon/SPARK-47066-followup.

Lead-authored-by: Hyukjin Kwon 
Co-authored-by: Hyukjin Kwon 
Signed-off-by: Hyukjin Kwon 
---
 .github/workflows/build_and_test.yml | 78 ++--
 1 file changed, 13 insertions(+), 65 deletions(-)

diff --git a/.github/workflows/build_and_test.yml 
b/.github/workflows/build_and_test.yml
index bad34fd746ba..eda86d1066fb 100644
--- a/.github/workflows/build_and_test.yml
+++ b/.github/workflows/build_and_test.yml
@@ -82,7 +82,7 @@ jobs:
   sparkr=`./dev/is-changed.py -m sparkr`
   tpcds=`./dev/is-changed.py -m sql`
   docker=`./dev/is-changed.py -m docker-integration-tests`
-  # 'build' and 'java-other-versions' are always true for now.
+  # 'build' and 'maven-build' are always true for now.
   # It does not save significant time and most of PRs trigger the 
build.
   precondition="
 {
@@ -91,7 +91,7 @@ jobs:
   \"sparkr\": \"$sparkr\",
   \"tpcds-1g\": \"$tpcds\",
   \"docker-integration-tests\": \"$docker\",
-  \"java-other-versions\": \"true\",
+  \"maven-build\": \"true\",
   \"lint\" : \"true\",
   \"k8s-integration-tests\" : \"true\",
   \"buf\" : \"true\",
@@ -788,17 +788,21 @@ jobs:
 path: site.tar.bz2
 retention-days: 1
 
-  java-other-versions:
+  maven-build:
 needs: precondition
-if: fromJson(needs.precondition.outputs.required).java-other-versions == 
'true'
-name: Java ${{ matrix.java }} build with Maven
+if: fromJson(needs.precondition.outputs.required).maven-build == 'true'
+name: Java ${{ matrix.java }} build with Maven (${{ matrix.os }})
 strategy:
   fail-fast: false
   matrix:
-java:
-  - 17
-  - 21
-runs-on: ubuntu-22.04
+include:
+  - java: 17
+os: ubuntu-22.04 
+  - java: 21
+os: ubuntu-22.04 
+  - java: 21
+os: macos-14 
+runs-on: ${{ matrix.os }}
 timeout-minutes: 300
 steps:
 - name: Checkout Spark repository
@@ -845,62 +849,6 @@ jobs:
 ./build/mvn $MAVEN_CLI_OPTS -DskipTests -Pyarn -Pkubernetes -Pvolcano 
-Phive -Phive-thriftserver -Phadoop-cloud -Djava.version=${JAVA_VERSION/-ea} 
install
 rm -rf ~/.m2/repository/org/apache/spark
 
-  apple-silicon:
-needs: precondition
-if: fromJson(needs.precondition.outputs.required).build == 'true'
-name: Apple Silicon build with Maven
-strategy:
-  fail-fast: false
-  matrix:
-java:
-  - 21
-runs-on: macos-14
-timeout-minutes: 300
-steps:
-- name: Checkout Spark repository
-  uses: actions/checkout@v4
-  with:
-fetch-depth: 0
-repository: apache/spark
-ref: ${{ inputs.branch }}
-- name: Sync the current branch with the latest in Apache Spark
-  if: github.repository != 'apache/spark'
-  run: |
-git fetch https://github.com/$GITHUB_REPOSITORY.git 
${GITHUB_REF#refs/heads/}
-git -c user.name='Apache Spark Test Account' -c 
user.email='sparktest...@gmail.com' merge --no-commit --progress --squash 
FETCH_HEAD
-git -c user.name='Apache Spark Test Account' -c 
user.email='sparktest...@gmail.com' commit -m "Merged commit" --allow-empty
-- name: Cache Scala, SBT and Maven
-  uses: actions/cache@v4
-  with:
-path: |
-  build/apache-maven-*
-  build/scala-*
-  build/*.jar
-  ~/.sbt
-key: build-${{ hashFiles('**/pom.xml', 'project/build.properties', 
'build/mvn', 'build/sbt', 'build/sbt-launch-lib.bash', 
'build/spark-build-info') }}
-restore-keys: |
-  apple-silicon-build-
-- name: Cache Maven local repository
-  uses: 

(spark) branch master updated: [SPARK-47083][BUILD] Upgrade `commons-codec` to 1.16.1

2024-02-18 Thread dongjoon
This is an automated email from the ASF dual-hosted git repository.

dongjoon pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new b61a00c26f0b [SPARK-47083][BUILD] Upgrade `commons-codec` to 1.16.1
b61a00c26f0b is described below

commit b61a00c26f0b738ea0f5603dd04fe605e0578ea3
Author: panbingkun 
AuthorDate: Sun Feb 18 18:56:46 2024 -0800

[SPARK-47083][BUILD] Upgrade `commons-codec` to 1.16.1

### What changes were proposed in this pull request?
The pr aims to upgrade `commons-codec` from `1.16.0` to `1.16.1`.

### Why are the changes needed?
1.The new version brings some bug fixed, eg:
- Fix possible IndexOutOfBoundException in PhoneticEngine.encode method 
#223. Fixes [CODEC-315](https://issues.apache.org/jira/browse/CODEC-315)
- Fix possible IndexOutOfBoundsException in 
PercentCodec.insertAlwaysEncodeChars() method #222. Fixes 
[CODEC-314](https://issues.apache.org/jira/browse/CODEC-314).

2.The full release notes:

https://commons.apache.org/proper/commons-codec/changes-report.html#a1.16.1

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
Pass GA.

### Was this patch authored or co-authored using generative AI tooling?
No.

Closes #45152 from panbingkun/SPARK-47083.

Authored-by: panbingkun 
Signed-off-by: Dongjoon Hyun 
---
 dev/deps/spark-deps-hadoop-3-hive-2.3 | 2 +-
 pom.xml   | 2 +-
 2 files changed, 2 insertions(+), 2 deletions(-)

diff --git a/dev/deps/spark-deps-hadoop-3-hive-2.3 
b/dev/deps/spark-deps-hadoop-3-hive-2.3
index 5aabe0e4aef1..dbbddbc54c11 100644
--- a/dev/deps/spark-deps-hadoop-3-hive-2.3
+++ b/dev/deps/spark-deps-hadoop-3-hive-2.3
@@ -36,7 +36,7 @@ cats-kernel_2.13/2.8.0//cats-kernel_2.13-2.8.0.jar
 chill-java/0.10.0//chill-java-0.10.0.jar
 chill_2.13/0.10.0//chill_2.13-0.10.0.jar
 commons-cli/1.6.0//commons-cli-1.6.0.jar
-commons-codec/1.16.0//commons-codec-1.16.0.jar
+commons-codec/1.16.1//commons-codec-1.16.1.jar
 commons-collections/3.2.2//commons-collections-3.2.2.jar
 commons-collections4/4.4//commons-collections4-4.4.jar
 commons-compiler/3.1.9//commons-compiler-3.1.9.jar
diff --git a/pom.xml b/pom.xml
index 64931dd73282..410cefb0e8d9 100644
--- a/pom.xml
+++ b/pom.xml
@@ -191,7 +191,7 @@
 3.0.2
 1.1.10.5
 3.0.3
-1.16.0
+1.16.1
 1.25.0
 2.15.1
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



(spark) branch master updated: [SPARK-47084][BUILD] Upgrade joda-time to 2.12.7

2024-02-18 Thread yangjie01
This is an automated email from the ASF dual-hosted git repository.

yangjie01 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 85108b0cb605 [SPARK-47084][BUILD] Upgrade joda-time to 2.12.7
85108b0cb605 is described below

commit 85108b0cb6059e9a5301b63ab266084defd0ddf2
Author: panbingkun 
AuthorDate: Mon Feb 19 10:15:37 2024 +0800

[SPARK-47084][BUILD] Upgrade joda-time to 2.12.7

### What changes were proposed in this pull request?
The pr aims to upgrade `joda-time` from `2.12.6` to `2.12.7`.

### Why are the changes needed?
The version `DateTimeZone` data updated to version `2024agtz`.
The full release notes: 
https://www.joda.org/joda-time/changes-report.html#a2.12.7

### Does this PR introduce _any_ user-facing change?
No.

### How was this patch tested?
Pass GA.

### Was this patch authored or co-authored using generative AI tooling?
No.

Closes #45153 from panbingkun/SPARK-47084.

Authored-by: panbingkun 
Signed-off-by: yangjie01 
---
 dev/deps/spark-deps-hadoop-3-hive-2.3 | 2 +-
 pom.xml   | 2 +-
 2 files changed, 2 insertions(+), 2 deletions(-)

diff --git a/dev/deps/spark-deps-hadoop-3-hive-2.3 
b/dev/deps/spark-deps-hadoop-3-hive-2.3
index 0b619a249e96..5aabe0e4aef1 100644
--- a/dev/deps/spark-deps-hadoop-3-hive-2.3
+++ b/dev/deps/spark-deps-hadoop-3-hive-2.3
@@ -139,7 +139,7 @@ jetty-util/10.0.19//jetty-util-10.0.19.jar
 jline/2.14.6//jline-2.14.6.jar
 jline/3.22.0//jline-3.22.0.jar
 jna/5.13.0//jna-5.13.0.jar
-joda-time/2.12.6//joda-time-2.12.6.jar
+joda-time/2.12.7//joda-time-2.12.7.jar
 jodd-core/3.5.2//jodd-core-3.5.2.jar
 jpam/1.1//jpam-1.1.jar
 json/1.8//json-1.8.jar
diff --git a/pom.xml b/pom.xml
index a14f2d255a90..64931dd73282 100644
--- a/pom.xml
+++ b/pom.xml
@@ -208,7 +208,7 @@
   Because it transitions Jakarta REST API from javax to jakarta package.
 -->
 2.41
-2.12.6
+2.12.7
 3.5.2
 3.0.0
 0.12.0


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



(spark) branch master updated: [MINOR][INFRA][DOCS] Remove undated comment in build_and_test.yml

2024-02-18 Thread yangjie01
This is an automated email from the ASF dual-hosted git repository.

yangjie01 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
 new 7c60fe21a29d [MINOR][INFRA][DOCS] Remove undated comment in 
build_and_test.yml
7c60fe21a29d is described below

commit 7c60fe21a29dd852de01da214c84e6a3deb38e31
Author: Hyukjin Kwon 
AuthorDate: Mon Feb 19 10:13:52 2024 +0800

[MINOR][INFRA][DOCS] Remove undated comment in build_and_test.yml

### What changes were proposed in this pull request?

This PR removes outdated comment. We don't use branch-3.3 anymore

### Why are the changes needed?

To remove obsolete information

### Does this PR introduce _any_ user-facing change?

No, dev-only.

### How was this patch tested?

N/A

### Was this patch authored or co-authored using generative AI tooling?

No.

Closes #45149 from HyukjinKwon/minor-ga.

Authored-by: Hyukjin Kwon 
Signed-off-by: yangjie01 
---
 .github/workflows/build_and_test.yml | 1 -
 1 file changed, 1 deletion(-)

diff --git a/.github/workflows/build_and_test.yml 
b/.github/workflows/build_and_test.yml
index c578d5079be8..bad34fd746ba 100644
--- a/.github/workflows/build_and_test.yml
+++ b/.github/workflows/build_and_test.yml
@@ -285,7 +285,6 @@ jobs:
   infra-image:
 name: "Base image build"
 needs: precondition
-# Currently, enable docker build from cache for `master` and branch (since 
3.4) jobs
 if: >-
   fromJson(needs.precondition.outputs.required).pyspark == 'true' ||
   fromJson(needs.precondition.outputs.required).lint == 'true' ||


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org